Try our new research platform with insights from 80,000+ expert users
reviewer1631958 - PeerSpot reviewer
Maintenance Manager at a transportation company with 10,001+ employees
Real User
We have seen quicker file transfers with more visibility and stability
Pros and Cons
  • "If they have ad hoc requirements, then they can theoretically schedule their own file transfers with the Self Service. We are trying to push as much work back to the customers or developers that have that requirement, because they prefer to help themselves, if possible. We try empowering them and enabling them through Control-M, especially for file transfers, because it is a much broader base of the business then just with batch scheduling. Typically, with SAP batch scheduling, it would work with dedicated teams. With file transfers, the entire business is involved. There are business users, end users, etc. It definitely needs to be as simple as possible and as managed as well as possible. They need to manage it themselves, if possible, because our team is not growing but the number of customers, applications, and jobs are growing. We need to hand back some of the responsibility to the customer for them to resolve and action it."
  • "The high availability that comes from BMC with its supplied Postgres database is very limited. Even using your customer-supplied Postgres database is problematic. We have engaged with them regarding this, but it is difficult. My company doesn't want to do this and BMC doesn't want to do that. We just need to find some middle ground to get the proper high availability. We're also moving away, like the rest of the world, from the more expensive offerings, like Oracle. We are trying to use Postgres, which is free. The stability is good. It is just that the high availability configuration is not ideal. It could be better."

What is our primary use case?

We schedule the majority of our SAP jobs Control-M. We do that globally for all the production plants. We have tens of thousands of SAP jobs and managed file transfer.

SAP batch and managed file transfer are critical processes that we have automated. We are in the process of replacing Connect:Direct and SecureTransport, the legacy file transfer solution, with Managed File Transfer (MFT). That is on the global scale. 

The Control-M for Informatica is gaining a lot of popularity, primarily in the financial side of the business. They have a lot of security restrictions that make their jobs very difficult. Also, there are cost issues for Informatica, e.g., anytime they execute a workflow in Informatica, they get billed for it. We are adapting the solution to not scrum the workflow every half an hour or hour because they pay for it, but only when it is needed. Therefore, we can do a database query and check if there are new records that need to be processed. If there are no records to be processed, then depending on that output, we either run the Informatica job or leave it and check again for maybe half an hour. We are optimizing, saving money for the customers and ourselves, while reducing the number of executions, jobs, etc.

We are using on-premises. We have been for many years. We are aware of the new Helix offering, which is a SaaS/cloud offering from BMC, but it is not really ready for enterprise yet, not at our scale. We are doing some cloud, though not the Helix offering. I have installations in the cloud using Azure and AWS. We are not fully functioning there yet. We are waiting for the demand, but we are aware of the cloud opportunities and making use of them.

We have been busy upgrading to version 9.0.20 Fix Pack 100 but our production environment is still on 9.0.19 Fix Pack 200.

How has it helped my organization?

We use Control-M as part of our DevOps automation toolchains and leverage its “as-code” interfaces for developers. We have found that a lot of the new customers who are developing for cloud prefer to use the API and would like to test for themselves. That is really where Jobs-as-Code comes in. They can test and fail quickly the agile way. We definitely have some customers who are using that.

We have seen quicker file transfers with more visibility and stability. Because data transfers are part of the Control-M tool, they form as part of the normal workflow. We see the value in that.

If they have ad hoc requirements, then they can theoretically schedule their own file transfers with the Self Service. We are trying to push as much work back to the customers or developers that have that requirement, because they prefer to help themselves, if possible. We try empowering them and enabling them through Control-M, especially for file transfers, because it is a much broader base of the business then just with batch scheduling. Typically, with SAP batch scheduling, it would work with dedicated teams. With file transfers, the entire business is involved. There are business users, end users, etc. It definitely needs to be as simple as possible and as managed as well as possible. They need to manage it themselves, if possible, because our team is not growing but the number of customers, applications, and jobs are growing. We need to hand back some of the responsibility to the customer for them to resolve and action it.

What is most valuable?

A new feature, which we deployed about two years ago, is the Managed File Transfer (MFT). We also use Managed File Transfer Enterprise (MFTE) for external transfers of our biggest use cases. 

Another valuable feature would definitely be the MFT dashboard that is now available in Control-M natively. It is easy to just search for jobs, files, etc. Instead of the customers contacting us to find out what happened, when it happened, and why it happened, they are able to service themselves. This allows us to cut down on operational staff, costs, and time because customers can manage it themselves to a degree.

The most valuable feature is definitely the Self Service. A couple of years ago, it was available, but not with the features that it is today. There wasn't really uptake on it, although it was available. We have seen a steady growth in the number of users using it and what they are using it to do. They are using Self Service to schedule by themselves and do monitoring by themselves. They interact with their schedules. Also, the performance of Self Service is very user-friendly and more accessible. That is one of the features that we use a lot lately.

The reporting has definitely improved over the years. We are definitely doing more of that as well. We are definitely seeing more value in reporting on the batch schedules, optimizing it and seeing if we can cut costs. 

What needs improvement?

The reporting has improved. It is not where it should be yet, but we have seen improvements. The biggest thing for me is the restrictions regarding templates for reporting. You can't create your report with your own parameters. We have a meeting weekly with BMC and our customer lifecycle architect, and this comes up quite frequently. We have been privileged enough to do work with the developers. They are aware of the requirements regarding reporting and what our customers are asking for.

What I found lately about the YouTube videos, specifically, is that they are very simple. Usually, when I watch a video, I would read the manual, instructions, etc. to see if I understand it. I would hope that the interactive sessions, Q&As, or videos could be used to handle more complex issues of what they're discussing. An example would be the LDAP authentication for the Enterprise Manager. They would typically just go through the steps that are in the documentation. What people typically looking at those videos are looking for is how to do the more complex setup, doing it with SSL and distributed Active Directory data mines. Things that are not documented. I find those videos helpful for somebody who is too lazy to read the manual. I expect them to handle more than what is available in the documentation and the more complex situations.

The high availability that comes from BMC with its supplied Postgres database is very limited. Even using your customer-supplied Postgres database is problematic. We have engaged with them regarding this, but it is difficult. My company doesn't want to do this and BMC doesn't want to do that. We just need to find some middle ground to get the proper high availability.
We're also moving away, like the rest of the world, from the more expensive offerings, like Oracle. We are trying to use Postgres, which is free. The stability is good. It is just that the high availability configuration is not ideal. It could be better.

Buyer's Guide
Control-M
June 2025
Learn what your peers think about Control-M. Get advice and tips from experienced pros sharing their opinions. Updated: June 2025.
860,592 professionals have used our research since 2012.

For how long have I used the solution?

I have been using Control-M for 12 years.

What do I think about the stability of the solution?

Control-M is really stable. We have seen that throughout the years. I have had customers who have been running version 6.3 for seven years after support stopped. It has been running for three years straight, without a reboot or restart, doing its job. We have actually had issues with customers who don't want to upgrade. They have said, "This stuff is working perfectly. Just leave it alone because it just doesn't go down." 

We have a saying in our department as well. When somebody says there is a problem, we say, "It's not Control-M. Check everything else. Check the server, network, and database. It's not Control-M." 99 out of 100 times, we are right. It is either infrastructure or something else, but it is not Control-M.

What do I think about the scalability of the solution?

I have never run into any problems scaling, either vertically or horizontally, with Control-M. In each version, it just gets better. I am really happy with that.

We were one of probably the first companies who bought MFTE, and it was not ready yet. It didn't scale properly. It didn't offer the functionality that the competing tools that we were currently using had. It's grown tremendously because of our input and feedback directly to the developers and BMC. I'm not complaining about it, but it put us back a bit. We have learned not to be a very early adopter. We have seen the same with the cloud. Everybody wants to jump on the cloud, but nobody knows why. They just want to do Cloud. We've made a substantial investment with MFTE. It was a couple of hundred thousand euros, and it was not ready yet for our enterprise requirements.

Our monitoring team who does 24/7 monitoring. They handle the alerts. They check their job flows. They make sure escalations are going through. If tickets need to be logged, make sure that gets done. They also interact with ad hoc requests from customers. 

There is the scheduling team who does the job definitions, updates, etc. 

There is the administration team, which I'm part of, with administrators who look after the infrastructure, Enterprise Manager, servers, agents, gateways, etc. Recently, we also have a dedicated MFT team that only looks after MFT because of the huge number of customers, requests, and requirements.

Other customers who use it are really all across the board. We had a presentation last week to our bigger department that is worldwide, but which we are a part of in South Africa. We have noticed about 52 main departments, then the sub-departments, between them. A lot of them sit right across the enterprise. Typically, the most active users would be SAP users who checks for output on the jobs running on Control-M. It is just 10 times easier to do it in Control-M than on SAP itself. We also manage to keep the output for longer than SAP. What they can't find on SAP after seven or 14 days, they can usually find with us, e.g., outputs for the jobs or logs. 

There are the MFT users who love being able to see each morning that their file was transferred, how long it took, and how big the file was. A lot of self-service users are using the Self Service function. Team leads and operational staff use it most.

How are customer service and support?

I love support and the support people. It is very good. Because we are quite a mature customer and the whole team has a lot of experience (sometimes more than the support people), if they don't realize the seriousness of the situation, then we would not escalate but just to make our customer lifecycle architect aware by saying, "We are not feeling this case is getting the required personnel on it. We need somebody more senior. We don't have time to cover the basics that the first line support is trying to deal with. We've been over that." Overall, I would rate the technical support as nine out of 10.

Which solution did I use previously and why did I switch?

Previously, we used a big SAP solution, which was not a commercial, and specifically designed for our company.

We have recently taken over a mainframe migration as well as the scheduling was on TWS, which is IBM's scheduling software on the mainframe z/OS. We moved that all over to Control-M. That was a combination of SAP jobs, Informatica jobs, database jobs, and normal script jobs. So, we use a bit of everything. We have also used the automation API a lot for interfacing with Control-M and other middleware tools, but primarily it is SAP and file transfer.

We use Control-M to integrate file transfers within our application workflows. It integrates with the tools that we are replacing, i.e., Connect:Direct, which is quite a legacy tool, and our old IBM tool, which we have been using for more than 15 years and has no visibility. With Control-M, you get visibility on your file transfers and how it mostly interacts with your batch schedule. Something gets created, it's sent over, and then it gets processed. Control-M has already been part of the executing, extracting, import, or processing. Now, with the file transfer, customers can see the entire workflow from the data being generated, transferred, and processed. This resolves a lot of complexities because you used to need to contact three different teams to find out if the file arrived and was processed. One tool does all of that now.

There isn't a lot of new functionality that our previous tools didn't have. It is just re-consolidating all the tools that we need into a single one. That makes it much simpler. There is one team to contact globally for file transfers, and that makes it easy. It provides visibility with its Self Service that wasn't available with Connect:Direct or SecureTransport. Our customers are quite happy to have that. We can also provide reports. 

SecureTransport competes with MFTE. There isn't a conversion tool for that yet. Connect:Direct simply provides the means for a conversion tool, but it gets integrated into scripts and applications. It's very difficult to migrate or extract that data.

How was the initial setup?

The initial setup is straightforward. It changed a lot over the years as well, but in the nicest way. You have minimal downtime with the upgrades on Enterprise Manager as well as the Control-M servers. A lot of preparation is done before the tool is shut down for the upgrade. Our downtime used to be at least an hour for upgrades or migrations. That has typically come down to 10, 15, or 20 minutes, depending on the size of the server. It is definitely more stable and understandable.

We have also noticed that the exception handling is much better if there are issues. We don't get that many surprises. The errors are understandable. The agent upgrades have zero downtime, so that is just amazing. All the patching and maintenance is centralized. We have migrated our development and integration environments to 9.0.20 in the last month or two. That went very smoothly. We will start with production next week. We have been through this quite a number of times. We came from version 7 to version 9 to versions 9.0.19 and 9.0.20. We do all the upgrades in-house.

What about the implementation team?

We do it all ourselves. If we get stuck, we would contact BMC. At my previous job, we were a partner for BMC in South Africa, and I was on the support side for BMC. It is only we need to open tickets for bugs or problems that we contact BMC. Typically, upgrades and migrations, we handle those in-house.

There are three people full-time on the administrative side. We have a global setup: Europe, Mexico, America, Africa, and China. We have tons of virtual machines and hundreds and hundreds of agents, and even more that we might host.

What was our ROI?

I know we have already budgeted for more tasks. The company is very happy with the performance of our teams, specifically the South African team. We are really doing more with good tools and less people. There is definitely a return on investment, just from the stability and visibility which has improved a lot.

On the effort side, we have definitely seen a lot of savings. We have some bigger projects that are automating the schedule and removing human intervention. These have reduced department staff/headcount, by about 50%, when we were able to automate the batch side of it, because also our department offers monitoring and operations as part of our service. We have a dedicated monitoring team. Whatever runs in Control-M, that is monitored by us and escalated, if needed. 

Departments now have multiple scheduling tools between the mainframe, distributed systems, and cloud. Control-M brings all of that, e.g., we have it on a single pane of glass so we can see the exact execution on the mainframe, the execution on the line, and the execution in the cloud. This is instead of using three or four different tools. Therefore, the complexity of batch monitoring and scheduling has decreased as well with the standardization of Control-M. That is definitely one of the big advantages that we have seen.

What's my experience with pricing, setup cost, and licensing?

It is expensive. We have a lot of customers who complained initially about the costs. Because it's not just the licensing, unfortunately. It's the infrastructure, salaries, etc. I like the licensing model. It is pretty straightforward. We are on the task license. I know that we have some really good discounts. Our BMC account manager makes sure that we stay below the license count as well as checking for growth. Overall, it's good. The licensing is simple enough for me. It is a bit expensive. Especially with the cloud coming in, we might see the licensing change in the future, but I'm guessing.

This is now from my previous years as support for banks and big companies. If it's not enterprise scale, I find that it's too expensive for smaller companies. You really have to be quite big and need to have a dedicated support staff to run it, then you'll be fine. What we've seen at smaller companies, it's too expensive because they want to automate everything. Now, stuff that can literally run once a day for the rest of their lives is costing them $3 a job a day. It becomes too expensive, eventually. They are not seeing the return on investment because it's not business critical. Nobody is going to die or they're going to lose money if that job didn't run exactly at 11 minutes past 4:00. It's definitely for bigger enterprise companies, especially banks or healthcare providers. We have had an instance where Control-M was unavailable due to external factors for 20 minutes and there was a loss of almost a million euros because the solution involved logistics. 

Which other solutions did I evaluate?

We have done the usual crontab migration. Everything is in crontab or Windows Scheduler. Typically, we end up with a migration, even if it's from a known tool, where we end by exporting it into Excel and converting it into job definitions with a script. We have been involved in that, but nothing using BMC tools.

When I joined the company, I first supported them through the local partner. Because we have such a vast array of scheduling tools, they went through a PoC and business case. We evaluated three or four tools, where BMC Control-M was one. Quite soon, because the company was already using Control-M in Africa and China, they were looking for global solutions to see if it really could create change.  

What it came down to was ease of use, enterprise capability, and BMC was already in the company with ITSM and a couple of other products as well. They had a good relationship with us. We consulted with other customers who have used it as well as references because it was expensive. It was definitely the most expensive solution then, out of the four. However, we didn't want to go five years down the line and then have to change again because of issues.

What other advice do I have?

We have had a very good run with Control-M. I love it.

With the move to big data and especially with our AWS Cloud presence, we have a data lake. We are in discussions with the analytics teams about how they can utilize Control-M in the cloud for analytics, big data, etc. However, at the moment, it is not a big deal.

What we have found with the Jobs-as-Code is that customers need to understand Control-M better, how the scheduling works, the knowledge around it, its conditions, etc. It took some time for the developers to get used to Control-M, then Jobs-as-Code. They are now confident with it. We are presenting twice weekly. We have an open forum for interested parties about Control-M or our department, Enterprise Scheduling and File Transfer, where we have a dedicated session about Jobs-as-Code. If there are questions about how other departments are doing it, if there is a better way to do it, if they are able to save on the number of jobs, can we make them rerun, or instead of creating 10 jobs, can it be done with five jobs? So, there is not a lot going from Jobs-as-Code directly into production, but we have a couple of parties, especially on the cloud front, who are very interested in it.

The solution is enterprise scale. Also, if you want to integrate all your applications into one view and offer all the functionality across the board, such as file transfer, scheduling, cloud, and on-prem, then you can create your own application integrations to integrate with applications that's not supported currently by BMC, like APIs. For top 100 enterprises, there isn't another better tool on the market for enterprise.

I would rate it as a nine out of 10.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
reviewer1657833 - PeerSpot reviewer
AVP - Systems Engineer at a financial services firm with 10,001+ employees
Real User
Allows us to integrate file transfers more readily, resolve issues quickly, and orchestrate a diverse landscape of vendor products
Pros and Cons
  • "The File Transfer component is quite valuable. The integration with products such as Informatica and SAP are very valuable to us as well. Rather than having to build our own interface into those products, we can use the ones that come out of the box. The integration with databases is valuable as well. We use database jobs quite a bit."
  • "A lot of the areas of improvement revolve around Automation API because that area is constantly evolving. It is constantly changing, and it is constantly being updated. There are some bugs that are introduced from one version to the next. So, the regression testing doesn't seem to capture some of the bugs that have been fixed in prior versions, and those bugs are then reintroduced in later versions."

What is our primary use case?

Control-M supports a lot of business processes. It supports some of the HR functions. I don't know if payroll is directly supported, but we do run jobs through PeopleSoft, which obviously impacts HR. Recently, we've started using the SAP module. So, we're making a transition from PeopleSoft to SAP, and I also see some payroll functions happening there.

How has it helped my organization?

We use Control-M to orchestrate a diverse landscape of vendor products such as Pega, MuleSoft, etc. File transfers and data feeds fetching are quite important for us. So, a lot of data processing happens through Control-M.

Control-M provides us with a unified view where we can easily define, orchestrate, and monitor all of our application workflows and data pipelines. Of course, such a diverse landscape requires you to make the effort to utilize Control-M to tie everything together or to act as the glue. Once you do that, everything is clearly defined, and you can view these disparate systems using one unified pane. If you don't define it correctly, then obviously Control-M won't have that insight, and so you'll have to go to multiple locations to go look at your job statuses.

We use its web interface. It is primarily for the application support teams to go monitor their own jobs. The jobs defined within Control-M are tightly controlled by a specific group of people. There are also people who need access to view that the jobs were completed successfully or why the jobs may have failed. These people are given access through Control-M web to view and monitor the jobs that they support or the applications they support. They're usually able to log on without having to install any client on their personal workstations. So, it's quite convenient. We have not implemented its mobile interface.

The integrated file transfers with our application workflows have certainly sped up our business service delivery by 80%. It has allowed the business to integrate file transfers more readily. Prior to utilizing the Control-M module, people had to write their own file transfer scripts in a scripting language of their choice to vary degrees of effectiveness. With the integrated File Transfer solution within Control-M, there is a standardized way of performing file transfers along with the capability of file watching and grabbing the file names that were transferred, making it much more versatile.

Control-M can immediately report when a job fails. If you have proper monitoring in place, you're notified immediately when your business flows are impacted. In the past, when you run jobs using Cron or just wrote shell scripts, you're really left in the dark because they don't necessarily report even from within Control-M. Implementing Control-M has made the business realize how critical and important it is to have proper error coding within the scripts that they schedule. If the scripts don't necessarily report any errors or redirect the system output into log files, when a job fails, there is no way to detect that.

We've automated many time-consuming business reports and other things that were very manual and took a tremendous amount of manhours. We've also automated a lot of maintenance using Control-M. We've integrated with Ansible Tower. So, we now are able to run Ansible playbooks and Ansible job templates. With the scheduling capability and the multitude of integrations that Control-M offers, it really acts as the unifying glue and as a communicator and orchestrator across the enterprise. With Ansible Tower, you can run a number of playbooks through it to perform patching and reboots and whatever maintenance that the infrastructure teams require, but you can't really do it when the business is still operating, or you can't do it when that business is operating, but you could do it for another business that's not operating at the moment. It is very hard to coordinate that without knowing which lines of business have jobs running or things like that. With Control-M, you can see that and you can actually enact workload policies to put jobs on hold prior to running Ansible playbooks. Once your Ansible playbook is complete, you can release the jobs again by deactivating the workload policies. So, it makes those processes very streamlined.

We do use the Role-Based Administration feature. We have been allowing other groups to gain more control over their agents so that they can define connection profiles, and they can do a little bit more on their side without inundating the main team with a lot of tasks. Everybody is happier. They can get things done faster, and they have immediate feedback and response because they're in control. The main Control-M team is not inundated with a lot of different requests from various teams to do a number of mechanical tasks. They don't get asked to create the connection profile for a database. People have all the information there, and they can do it themselves. They can define it in a way so that only they have access to it.

It has helped us to achieve faster issue resolution. Control-M reports on the error. It is easier to view the system output of that job. Whether it is an Informatica job, a scripted job, or a database job, it is easier to go in and view the issue and then troubleshoot from there. Most of the time, you can be running from the point of failure if the jobs aren't defined correctly. In a properly defined job, I would estimate that there is a 70% to 90% reduction in the meantime to resolution.

It has helped us by improving our service-level operations performance. We've built integration between Control-M and our ITSM, which is ServiceNow, and that has certainly allowed us to gain more visibility within our community through ServiceNow. Every time a production job fails, an incident ticket is cut, and that's highly visible. That needs to be escalated too, and there is a much more defined process to be able to resolve that issue. In the past, obviously, when you didn't have that level of visibility or that integration, there was always time lost in identifying what the issue is.

What is most valuable?

The File Transfer component is quite valuable. The integration with products such as Informatica and SAP is very valuable to us as well. Rather than having to build our own interface into those products, we can use the ones that come out of the box. The integration with databases is valuable as well. We use database jobs quite a bit. The file watcher component is also indispensable when integrating with other applications that generate files, instead of triggering a workflow based on time.

What needs improvement?

We have been experimenting with centralized connection profiles. There are some bugs to be worked out. So, we don't feel 100% comfortable with only using centralized connection profiles. We do have a mix of control on agents out there, which leads to some complications because earlier agents do not support centralized connection profiles.

A lot of the areas of improvement revolve around Automation API because that area is constantly evolving. It is constantly changing, and it is constantly being updated. There are some bugs that are introduced from one version to the next. So, the regression testing doesn't seem to capture some of the bugs that have been fixed in prior versions, and those bugs are then reintroduced in later versions. One particular example is that we were trying to use the Automation API to fetch a number of run ads users from the environment. The username had special characters and backspace characters because it was a Windows User ID. In the documentation, there is a documented workaround for that. However, that relied on two particular settings in the Tomcat web server. I later found out that these settings work out-of-the-box for version 9.0.19, but those two options were not included in the config file for 9.0.20. So, it led to a little bit of confusion and a lot of time trying to diagnose, both with support and the BMC community, what is the issue. Ultimately, we did resolve that, but that is time spent that really shouldn't have been spent. It had obviously been working in 9.0.19, and I don't know why that was missed in 9.0.20, but that's a primary example of an improvement that can happen.

We've also noticed that the Control-M agents themselves now run Java components. Over time, they tend to destabilize. It could be because garbage collection isn't happening, or something is not happening. We then realize that the agent is consuming quite a large amount of memory resources on the servers themselves. After recycling the agents and releasing that memory, things go back to normal, but there are times when the agent becomes unresponsive. The jobs get submitted, and nothing executes, but we don't know about it until somebody says, "Hey, but my job isn't running." When we look at it, it says Executing within the GUI, but there is no actual process running on the server. So, there is some disconnect there. There is no alerting function or the agent there that says, "Hey, I'm not responding." It is not showing up in the x alerts or anything like that.

The integrated guides have not been that helpful to us. I do find a lot of the how-to videos on the knowledge portal to be useful. However, there are some videos where the directions don't always match with some of the implementations. There are some typos here and there, but overall, those have been more helpful for us.

Its pricing and licensing could be a little bit better.  The regular Managed File Transfer piece, is a little overpriced, especially for folks who already have licensed Advanced File Transfer.

What I'm also noticing when I'm trying to recruit for Control-M positions is that the talent pool is quite small. There's not a whole lot of companies that utilize Control-M, and if they do, most people don't want to let their Control-M resources go if they're good. There is a high barrier of entry for most people to learn Control-M. There are Workbench, Automation API, and so forth mainly for developers to learn, but there are not a whole lot of resources out there for people to get more familiar with administering Control-M or things like that in terms of the technology or even awareness. So, it becomes very challenging to acquire new resources for that. A lot of the newer people coming out of college don't even know what is Control-M. If they do, they think of it as a batch scheduler, which is certainly not true in its current transformation.

Control-M is a very powerful enterprise tool, but the overall perception has not changed in the past five to six years that I've been working with Control-M. There's not much incentive for people to dive into that world. It is a very small community, and overall, the value of Control-M is not being showcased adequately, maybe at the C-level for corporations. I've had multiple conversations with other people and other companies who have already exit using Control-M. About 70% of the companies out there do not take full advantage of the capabilities in Control-M. So, that type of utilization really hampers and hinders the reputation of Control-M. That's because people then acquire this untrue concept that Control-M can only do X, Y, and Z, rather than the fact that Control-M can do so much more. I don't know if it needs a grassroots marketing movement or a top-down marketing movement, but this is what the perception is because that's what I'm hearing and that's what I'm seeing. For some of the challenges that I face working in Control-M, when I go back to my management and say, "Hey, I want to spend more money in this space," they're like, "Why? Can you justify it? This is what we see Control-M as it is. It's not going to bring us value in this area or that area." I have to go back and develop a new business case to say, "Hey, we need to upgrade to MFT enterprise or something like that." So, it definitely requires a lot more work convincing management in order to get all these components. In the past, we had to justify acquiring a workload change manager. We had to justify acquiring the workload archive. All of these bring benefits not only to our audit environment but also to the development environment, but the fact that we had to fight so hard to acquire these is challenging.

For how long have I used the solution?

I've been using Control-M for about eight years.

What do I think about the stability of the solution?

Version 9 was very stable. Once they started adding a lot of the newer Java components, the stability suffered. It seems to have gotten better in version 9.0.20, but that's could be my basic perception. 

We run a lot of database client jobs. There are some things that we've implemented that I understand can contribute to the agent instability. We sometimes extract a lot of database output and massage that output using other scripts. I've noticed there are certain things that you cannot do with it, or there are some things that contribute to the instability. For example, in the output scanning functionality, there certainly is a size limit. You probably don't want to scan anything too large because that's going to put a lot of resources on the environment.

In addition, there are times when the agent becomes unresponsive. The jobs get submitted, but nothing executes. There is no alerting function. These are the examples of instability that I've noticed. Overall, the main application itself, the EM, and the scheduler have been pretty stable.

What do I think about the scalability of the solution?

It is very scalable in terms of job execution. I haven't really explored scaling Control-M and the EM environment to a point where we have hundreds of users accessing it at a given time. That's because I don't have a hundred users who want to access that at a given time, but I do understand that you can distribute the web server more, and then have a load balancer to balance the load. I would think Control-M is a fairly scalable application.

In terms of its users, we have a lot of application support folks. We do have some developers who access Control-M mostly for the non-prod environments to execute and monitor their own jobs. There are some software engineers and operational engineers who are part of the application support teams that access Control-M. As for size or concurrent users, we have about 50 concurrent users at the max.

How are customer service and support?

I would probably give them a nine out of 10. For the most part, they're very helpful, but there's always an initial standard dialogue. For an issue, you have to collect from EM logs, agent logs, and so forth, and you submit that. Sometimes, we have done all the advanced work and submitted it, but they still come back and say, "Hey, we need the logs." It seems like that's a canned response without looking at the tickets.

How would you rate customer service and support?

Positive

Which solution did I use previously and why did I switch?

We've been with Control-M for quite a long time. We have not been using anything else in my history with this organization. 

I have not looked at anything recently. I am aware there are other application orchestration solutions out there, but I have not felt the need to go explore those options at the time.

How was the initial setup?

If you're deploying using out-of-the-box options, the process is fairly straightforward. If there is some customization that needs to happen, then the process can be complex, and the documentation does not cover some of those complexities.

For the most part, we are standard out of the box. We have run into some performance issues where we had to, later on, go in and maybe make some modifications. For example, we had to stand up different gateways for various purposes just because one singular gateway was not enough to take the load in particular because we had installed a workload archive, and that was just taking up a lot of resources. Other human users were not able to perform their actions because the archive user was consuming so much of the server's resources. So, there was a lot of tweaking there, and we had to basically break out and distribute some of the components.

In terms of implementation strategy or deployment plan for Control-M, the environment always had Control-M, and we just had to upgrade the Control-M environment. We've had Control-M in our environment for quite a long time, probably when it was still version 6. So, as we progressed through different versions, we obviously had to expand the environment and the platforms. We initially started off with Control-M on AIX, and we later moved to Control-M on Linux. As you go to Linux, obviously, there is planning for high availability and production environments, disaster recovery environments, and so forth. So, you have to plan for marrying a lot of the BMC Control-M components and identifying where a load balancer may be required, or DNS ALIAS is required so that you can quickly flip over in the event something happens. Then, of course, there is sizing for the environment in terms of how many jobs are running, how many executions are happening, and so forth. This is how we plan.

What about the implementation team?

We've used the AMIGO program, and then we've performed the upgrades ourselves.

For its day-to-day administration, we have a team of five people. They're administrators and schedulers.

What was our ROI?

Its return on investment is quite high, and that's mostly because we use so many of Control-M's capabilities. We also extend those capabilities. We write our own scripts to be able to integrate Control-M with so many other applications such as Automation Anywhere, Alteryx. We have also done vice versa. We have helped other teams develop their capabilities in integrating with the REST API and Control-M. So, the ROI is quite high for our use case, but based on the conversation with some of the community partners out there, their ROI is probably quite low because they're not making use of all these new features. I don't know if it is because they don't have the skillset to make use of these new features, or their management structure or process structure is hampering them from going out there. A lot of large companies I know like to maintain the status quo, and that's why they're slow to adapt and slow to move, which is going to hurt them in the long run, but in the meantime, it can hurt the adoption of Control-M as well.

What's my experience with pricing, setup cost, and licensing?

Its pricing and licensing could be a little bit better. Based on my experience and discussions with other existing customers, everybody feels that the regular Managed File Transfer piece, not the enterprise one, is a little overpriced, especially for folks who already have licensed Advanced File Transfer. We understand that Advanced File Transfer is going away and is going to be the end of life, and there is some additional functionality built into MFT, but the additional functionality does not really correlate with the huge price increase over what we're paying for AFT already. This has actually driven a lot of people to look for alternative solutions.

I know they are now moving more towards endpoint licensing or task-based licensing. In my eyes, the value of Control-M is the ability to break down jobs from monolithic scripts. You don't want to have to wrap everything up in one monolithic script and say, "Hey, I'm executing one task because I want to save money." That defeats the purpose of controlling, and that defeats the value of Control-M. By being able to take that monolithic script and break it down into the 10 most basic components, you can monitor each step. It is self-documenting because, within Control-M, you can see how the flow will work, and you can recover from any one of those 10 steps rather than having to rerun the monolithic script should something fail. That being said, the endpoint licensing does make more sense, but maybe pricing or things like that can be more forgiving.

Which other solutions did I evaluate?

N/A

What other advice do I have?

It is worth the time and money investment to learn more about Control-M. You should learn all the features of Control-M and really explore and test out the capabilities of Control-M. That's the only way people get comfortable with what Control-M can implement. A lot of people aren't aware of just how flexible a platform Control-M is, especially with all the new features that are being added via the Automation API. These features are helping to drive Control-M and things developed in Control-M more towards a microservices model.

We are just beginning to explore using Control-M as part of our DevOps automation toolchains and leverage its “as-code” interfaces for developers. Obviously, there is a little bit of a learning curve for developers as well in order to see the value of developing Jobs-as-Code. Currently, we're walking developers through it, and we're holding their hands a little bit in terms of developing Jobs-as-Code, but we are heading in that direction because it does provide artifacts that you can version control and change quickly and easily. You can redeploy much quicker than just having the jobs defined in the graphical user interface. Previously, when you had to modify it, you either did it via the GUI, or you exported it via XML and then modified those components. Once you get the developers closer to their job flows, then you can theoretically speed up the delivery of applications along with scheduled jobs.

I don't have a whole lot of experience with other scheduling orchestration environments, but from everything that I've heard while speaking with other colleagues, I would say Control-M ranks fairly high. I would rate it a nine out of 10. Control-M usually is the platform that people are moving to, not moving away from.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
Buyer's Guide
Control-M
June 2025
Learn what your peers think about Control-M. Get advice and tips from experienced pros sharing their opinions. Updated: June 2025.
860,592 professionals have used our research since 2012.
Director Information Technology at a insurance company with 1,001-5,000 employees
Real User
Enabled us to consolidate and streamline our development process, while building on existing skills
Pros and Cons
  • "We used Control-M's Python Client and cloud data service integrations with AWS and, as a feature, it was very customizable. It gave us a lot of flexibility for customizing whatever data maneuver we wanted to do within a pipeline."
  • "I would like to see them adopt more cloud. Most companies don't have a single cloud, meaning we have data sources that come from different cloud providers. That may have been solved already, but supporting Azure would be an improvement because companies tend not to have only AWS and GCP."

What is our primary use case?

Our use case was mainly about consolidating our data pipeline from different sources and doing some data transformations and changes. We needed to get data from different sources into a state where we could act on it into one consolidated data set.

How has it helped my organization?

It gave us the ability to consolidate a diverse set of solutions into one comprehensive solution that streamlined our development processes. It was straightforward to adopt and we could build on existing skills without having to have 10 solutions for 10 problems.

And when it came to creating actionable data, it gave us the ability to move faster and at scale. By adopting a solution like Control-M, we were able to scale and deliver faster data transformations and maneuvers, turning data into insights in a more efficient and scalable way.

The ability to deliver faster and at scale was important. Business and management always wanted us to deliver faster and bigger and we were able to do both with the solution that we implemented using Control-M. We were able to respond faster to changes and business needs, at scale. 

Having a feature-rich solution enabled us to aggregate all of our processes into it, and that made the overall execution, from a project and portfolio perspective, a lot more efficient.

We were also able to respond to audit requests, because it's centralized, in a much more efficient way.

What is most valuable?

There isn't a single feature that is most valuable, but if I had to choose one, it would be the rich ability it gave us for making customized scripts. That was probably the most unique feature set for our situation. We used Control-M's Python Client and cloud data service integrations with AWS and, as a feature, it was very customizable. It gave us a lot of flexibility for customizing whatever data maneuver we wanted to do within a pipeline.

The Python Client and cloud data service integrations have a rich set of features with flexibility. It did not require additional, crazy skills or experience to deal with it. It was a nice transition into enabling a data scientist to leverage existing skills to build those pipelines.

Creating, integrating, and automating data pipelines with Control-M was straightforward. It did require some knowledge and training, but compared to other solutions, it was a lot simpler. Working with data workflows, with the data-coding language integrated into Control-M, was straightforward. The level of difficulty was somewhere between "medium" and "easy." It was not that hard to leverage existing skills and knowledge within this specific feature.

The user interface for creating, monitoring, and ensuring delivery of files as part of the data pipeline was very actionable. It was almost self-explanatory. Somebody with basic user-interface experience could navigate the calls to action and the configuration that is required. It was well-designed.

What needs improvement?

I would like to see them adopt more cloud. Most companies don't have a single cloud, meaning we have data sources that come from different cloud providers. That may have been solved already, but supporting Azure would be an improvement because companies tend not to have only AWS and GCP.

For how long have I used the solution?

I used it for a couple of years.

What do I think about the stability of the solution?

It's fairly stable. I don't recall any specific issues. 

What do I think about the scalability of the solution?

It's fairly scalable. For our needs, it scaled very nicely.

We have a shared model where we have a centralized, shared service organization when it comes to data. Different people will use it, but it's centralized.

How are customer service and support?

We used other solutions from BMC as well, and their customer support was always great. I give them a 10 out of 10.

Training or a Knowledge Base were available or you could ask a question by submitting a ticket.

How would you rate customer service and support?

Positive

Which solution did I use previously and why did I switch?

We had DataStage from IBM and SSIS.

The switch was really about streamlining the process. We had other tools that only did partial processes or were not doing it with the speed and efficiency that we were looking for. We were looking for a solution that could streamline things and solve 90 percent of our data challenges.

What was our ROI?

The analysis that I saw validated that the ROI was within a couple of years.

What's my experience with pricing, setup cost, and licensing?

The pricing was competitive, from what I understand.

Which other solutions did I evaluate?

We looked at continuing to use the same solutions we had been using, and there were a couple of other cloud-based solutions that we evaluated. One of them was Matillion. The ease of use was one component of our decision, as was the flexibility of scripting with Python. Those were the key differentiators.

What other advice do I have?

For the on-prem solution, we had to do the patching and whatever was required by the vendor, but the cloud implementation was a model that was usable. The upgrades, changes, and patching are done directly by the vendor.

Control-M was a critical piece of the puzzle, to help us with all the data transformation and projects that we had to do. It was part of either one specific project or even a larger project that required that middle data transformation so that we could get to analytics or any other consumption of that data.

Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
Nagarajan Sankarammal - PeerSpot reviewer
Automation Architect at a tech vendor with 10,001+ employees
MSP
A highly capable, feature-rich solution with excellent third-party integrations
Pros and Cons
  • "The product has improved dramatically over the years; it offers a lot in terms of features and capabilities and integration with third-party tools. A wide range of models available with the product is critical in reducing manual and mundane work such as custom script writing. This saves significant amounts of time and, by association, money for the organization."
  • "Regarding product design and R&D, the DevOps pipeline could be improved with better capabilities and automation. API security and authentication is another area that could use improvement; users must have static credential passwords, which is a security concern."

What is our primary use case?

We use the solution to automate our business batches, workload processing, and some elements of our IT and system maintenance procedures and processes. These include sequential clips, programs, and workflows. We automate these and have them scheduled for regular execution. We needed an orchestrator, and Control-M fits our requirements well.

How has it helped my organization?

The availability of data and reports is vital, and the solution's capacity for timely processing and build generation improved considerably over time. As our operation grew, so did our use of Control-M, and there has never been a delay in the availability of data and reports, even with very high workloads. Eventually, we could also bring automated control over our back end. Control-M makes workflow orchestration simpler; it can deal with an impressive amount of transactions.

We realized the benefits of the solution a long time ago, and from time to time, there will be a situation that reminds us how valuable it is to us. Control-M is an overwhelmingly stable and steady product, free from issues and frequent disruptions. As is the case for any such tool, there are occasional bugs and fixes, but overall, it's a stable product and a fully integrated part of our operation.

What is most valuable?

The product has improved dramatically over the years; it offers a lot in terms of features and capabilities and integration with third-party tools. A wide range of models available with the product is critical in reducing manual and mundane work such as custom script writing. This saves significant amounts of time and, by association, money for the organization.

As a Control-M user for over 15 years, I see it as very straightforward to create, integrate, and automate data pipelines. Even in the beginning, when Control-M was more of a data architecture product, it was easy to pick up. I've seen multiple people adapting very well in terms of adapting and enabling the capabilities of the solution for business; it's straightforward. 

Ideally, agent lift modes of connectivity would be established on different platforms. We can get applications integrated directly with Control-M. That's a recent feature. There are ready-made platforms and plugins which allow us to see templates for workflow orchestration in third-party and custom in-house applications. It's a straightforward solution, and this is an area where Control-M excels.  

Our customers are pleased being with Control-M, despite some minor hiccups, which happen with any solution. They have been happy with the product for years, and it's an enterprise-wide batch workflow orchestration tool. That's how it is established in our organization and what our users are satisfied and familiar with. 

The process execution speed is excellent and has constantly improved over the years.

The bottom line is Control-M is a mission-critical solution, it's integral to our organization. 

What needs improvement?

Regarding product design and R&D, the DevOps pipeline could be improved with better capabilities and automation. API security and authentication is another area that could use improvement; users must have static credential passwords, which is a security concern.

The REST API supports FTP for file transfers, but we would like to see additional, more encrypted protocols and simplified file transfer encryption. Currently, the solution offers PGP encryption, which isn't the most straightforward. 

For how long have I used the solution?

We have been using the solution for around 15 years. 

What do I think about the stability of the solution?

The product is stable.

What do I think about the scalability of the solution?

The solution is scalable; we increased our usage over the years and plan to continue that. 

We have multiple teams at multiple geos and deployments; we're an enterprise-sized organization.

How are customer service and support?

Recently, there are some bugs with the product development, which necessitated R&D's involvement, which isn't ideal. We have fully integrated the solution into our production businesses, so any stability issues have a significant impact. There were cases where workarounds weren't provided quickly, with stubborn bugs needing environmental solutions.

How would you rate customer service and support?

Neutral

Which solution did I use previously and why did I switch?

I've worked with multiple other workload orchestration tools, including IMB Tivoli Workload Scheduler and a CA automation product. Control-M stands above the competitors in terms of stability. CA underwent an acquisition, leading to changes in product strategy and mergers with equivalent products like Automic, so Control-M was the surer option. It is also more robust and has greater system availability than the competitors.

How was the initial setup?

The initial deployment was complex, and this was by necessity. It is important to note that deployment is now more straightforward due to years of knowledge, experience, and newer features.

It took around two weeks to set up the Control-M infrastructure, and the process of bringing in business data and full adoption took place over years. It could be done faster, as in our case, there were other considerations involving budgeting, testing, and timelines. Setup of the initial infrastructure takes a few weeks, and then getting the platform running and configured can be done in a day or two. Further configuration and integration with LDAP and monitoring tools can take a little longer.

The solution can be managed and maintained by two or three staff members, but the number of staff involved in a deployment can vary significantly. It depends on the specific scenario and teams.

What was our ROI?

I would say we have a return on our investment; we have a vast amount of transactions and business automation implemented on a massive scale. We have adopted Control-M extensively, and it would be challenging to migrate to another solution in a reasonable time. We often look at alternatives but considering factors like timelines, resource availability, and team bandwidth, we keep coming back. The solution provides enormous value to our organization.

What's my experience with pricing, setup cost, and licensing?

The solution is not cheap, it comes with quite a hefty price tag. Control-M is the market leader, but we still want the price to be as friendly as possible. 

The solution comes with the base module and an additional one with a few extra plugins, which is helpful. 

Which other solutions did I evaluate?

We evaluate competitors yearly, but in terms of value for money, we always return to Control-M. We get an excellent return on our investment.

What other advice do I have?

I would rate this solution an eight out of ten.

I would advise any organization to do a proof of concept for their scenario before making a decision.

We don't currently use the Python Client, it's something we are planning to look into. We haven't started working on it, but we are in the review process to understand the client, and how it could fit into our operation.

The solution doesn't create new data as such, but it processes on top of the business data. 

We don't currently use the product for analytics, but we do plan to get Control-M data onto other systems for analytics and machine learning tasks.  

If we didn't have Control-M we would use an alternative solution. If there was a better one we would use that, or a product with a favorable cost and value proposition, which is a key factor.  

The tool always positively impacted our business, including our business service delivery speed. Over the years, there were one or two issues, but the vendor supporters could keep up. Some bugs required extensive development, and the support is excellent in this regard. They always have the right staff to assist during major productions or changes. Compared to before we had Control-M, it's as if we were previously traveling by foot, and now we've discovered the wheel. 

Regarding the audit preparation process, features like workload archiving come with an additional cost, which not all organizations can afford. I would instead maintain something locally on the system, but the solution is straightforward in terms of data necessities.

Control-M has to catch up in some areas, but it also offers specific capabilities and customization options. Application integration provides scope for exploration and deployment in custom developments. As a product supplier, BMC could focus on improving in areas indicated by their biggest customers. There is a lot of room for improvement.

File transfer support is Control-M's only significant limitation, as it only supports SSTP and STP transfers. Overall, other tools offer better security and file size in terms of file transfers. Therefore, the solution slows down when dealing with larger file sizes or a high volume of files, but it is sufficient for smaller organizations.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
Control-M Administrator at Cognizant
Real User
User-friendly GUI, responsive support, and the BIM feature helps us meet our SLAs
Pros and Cons
  • "BIM is helpful because we do not miss any SLAs, as we get to know the issue well in advance. It is the topmost service that has helped us provide better solutions for the business."
  • "The reporting functionality needs a lot of work. We have faced problems with different versions where we run the right report, but it gives us blank entries. Then, when we run the same report again, it gives the correct data."

What is our primary use case?

Our organization has multiple projects that use Control-M, and I support the banking domain. In the past, I have worked on projects for retail organizations and medical companies.

We have approximately 150 applications in our current project. These include Loanpower, erwin, and OpenLink.

How has it helped my organization?

With the use of Control-M, our SLAs are met more often. If there is an issue, we identify it in advance, before the problem occurs.

Control-M helps us in terms of automation because it has various scripts in different formats. We can run a Python program or a shell script, and these allow us to automate almost everything.

This product helps to secure our business because we can restrict users.

We have automated several critical processes with Control-M. One is used during patching, where we log in and type one command that will stop and start the services on all of the servers that we have. We have approximately 10 servers in production and five in non-production, so it's a lot of work to restart all of the servers. We also have automation that performs a health check. It runs every day at a scheduled time and will delete all jobs in production that are older than five days. Similarly, we have jobs that check to ensure certain conditions are being met and will check the various alerts that can occur.

Automating these processes has improved our business because every morning, we have to send a status update to show that the components are working. This is something that we used to do manually. We would log into CCM and check everything. Now, we have automated that using a script, wherein it sends the status email automatically to whichever business users request it. It has helped to reduce a lot of manual activity.

Control-M has definitely helped us to resolve issues faster. I estimate that the improvement is between 60% and 70%.

Our service-level operations performance has improved by 80% with the use of Control-M.

What is most valuable?

The GUI is very user-friendly. It provides us with a single view and we have everything in the same UI. This is very important because we don't spend a lot of time switching tabs or opening Control-M for different purposes. We have a single GUI open and it saves a lot of time.

Two really helpful features are Forecast and Business Impact Manager (BIM).

BIM is helpful because we do not miss any SLAs, as we get to know the issue well in advance. It is the topmost service that has helped us provide better solutions for the business.

Forecast is useful in terms of patching, etc, because whenever we are looking for downtime or any team is looking for downtime, it's easy for us to use Forecast to find it.

Self-service is helpful and our business users appreciate it because they don't have to have Control-M installed on their machine. They can log in using the web portal.

What needs improvement?

The reporting functionality needs a lot of work. We have faced problems with different versions where we run the right report, but it gives us blank entries. Then, when we run the same report again, it gives the correct data. We have spoken with Customer Care and some of the issues are fixed in the latest version, 9.20.

For how long have I used the solution?

I have been using Control-M for 11 years and my company has been using it for longer than that.

What do I think about the stability of the solution?

This is a pretty stable solution. We have not had any downtime.

A couple of times, the agent has gone down unexpectedly. However, in terms of the EM and server, it's pretty stable.

What do I think about the scalability of the solution?

Our organization is pretty big, with approximately 250,000 employees, and we have multiple projects that use Control-M. We have approximately 150 applications in our current project, and there are about 175 employees that are actively using Control-M. That is across three different countries.

It is easy to scale. It can handle a lot of job flows and it's easy to create multiple jobs to run at the same time. We are expanding in terms of jobs for the same application because they have a lot of upgrades going on at the application level. 

We are not planning to expand the number of applications in our project as of now. We do have requests, but it's a slow process. We can add perhaps five or six applications a year.

Overall, we have no problems in terms of scalability. 

How are customer service and technical support?

When we can't find a solution to an issue, we reach out to BMC customer support and they respond almost immediately. Overall, the technical support team is very good and I would rate them a nine or ten out of ten.

Which solution did I use previously and why did I switch?

We did not migrate to Control-M from a competing solution. Some of our clients, although not my current project, migrated to Control-M from different products. The reasons for changing products are the additional features available in Control-M, as well as the ease of use. Also, some people are more confident in the security that Control-M provides, compared to other tools on the market.

Personally, I started my career with Control-M and have been using it ever since.

In the company, we have a couple of clients who use IBM Tivoli Workload Scheduler (TWS), AutoSys, and Stonebranch. However, the majority of our clients use Control-M. The choice of solution stems from requirements and input from the client.

One of the reasons that some clients are not using Control-M is because of the cost. For a client with 5,000 or more jobs, they definitely implement Control-M. However, if they are running only 200 or 300 jobs in a small environment, there are other native tools available.

How was the initial setup?

I was not part of the implementation at my company but I have implemented several Control-M projects. The initial setup is straightforward.

First, we download the files from the BMC site and then start the installation. This involves running the setup files and if there is any error, you have knowledge base articles and you also have AMIGO support if you enroll in it.

The deployment can be completed in a day or two, including the Enterprise Manager (EM), servers, and agents. There are also conversion tools that are available to assist with creating jobs.

Our implementation strategy began with installing the Enterprise Manager first, and then the server, and then the agents. We would raise a support ticket so that whenever we had any issues, we could reach out to them.

I did not look at the interactive guides or videos that Control-M provides for reducing time to full productivity. I had all of the documentation handy but I did not refer to any of the videos.

What about the implementation team?

Our in-house team is responsible for deployment.

What's my experience with pricing, setup cost, and licensing?

Control-M is priced accordingly for larger environments. It is expensive for smaller environments with only a few hundred jobs running.

There are two different types of licenses available. The first is based on the number of jobs that we run per day, and the other is based on the number of agents that we install. My current project has a contract for five years. During the first two years, we are allowed to run any number of jobs using any number of agents. However, in the last three years, we have to stick to whatever is defined in the contract.

In past versions, BIM and Forecast were separate components that were available at an additional cost. Since version 9, however, everything is included and there are no costs in addition to the standard licensing fees.

Which other solutions did I evaluate?

For my current project, the client has always used Control-M.

What other advice do I have?

The latest version of Control-M is 9.20 but we are working with 9.18 because our client has certain servers where the OS is not compatible with 9.20. It is running on Linux machines and at this point, our client hasn't given us approval for the OS upgrade.

Our business users don't typically use Control-M. They have access to it but only use it when a critical chain is stuck and they want to check it themselves. They can use self-service for this, although most of the time, they don't.

An example of why they would use self-service is when a critical batch has failed and is stuck for a long time, and they want to see the approximate time that it will be completed. Also, during an audit, they can use self-service to see which users have certain access, such as production access or write permissions.

The Control-M users in our company have different roles. We have administrators, and we have people who specialize in migration. We also have people who look into scheduling and we have a team that just takes care of monitoring.

The number of people that we require for the day-to-day administration depends on the size of the project. In my current project, we have approximately 8,000 jobs actively running. We have approximately 17,000 configured. In our L1 team, we have eleven people, and we have eight members for each of our L2 and L3 teams.

We do not use the Control-M integrated file transfer capability in our workflows, although we do use the File Watcher feature. We have a tool from Axway called SecureTransport, where they handle the file transfer, but we can define this as part of a Control-M job.

The biggest lesson that I have learned from using Control-M is that anything can be automated. You can control various applications and it is simple to schedule jobs for products like SAP and databases.

My advice for anybody who is considering Control-M is that it has a wide variety of features compared to other tools. It is flexible, easy to use, and the web portal makes it simple for business users or application teams to access it without having to install it on a Windows server or a Citrix platform. 

I would rate this solution a nine out of ten.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor. The reviewer's company has a business relationship with this vendor other than being a customer: Partner
PeerSpot user
Subject Matter Expert at a consumer goods company with 10,001+ employees
Real User
Workflow dependencies work well, and automated audit reporting helps out sort out issues quickly
Pros and Cons
  • "Workload Archiving is a very good feature for us. It helps with our customer requirements in terms of reporting and auditing... Previously, when we didn't have any archive server, we managed the data in Control-M with man-made scripts, and we would pull the data for the last 365 days, or three or four months back. Since we installed the archiving, we have been able to pull the data, anytime and anywhere, with just one click."
  • "With the current version update, I'm not sure why we needed a separate database upgrade. Why not put it all in one package? Previously, you could do it either via a manual upgrade or an in-place upgrade but it wasn't separate."

What is our primary use case?

I am working with a beauty products company and we are dealing with supply chain issues. Most of the jobs in Control-M are through SAP.

Right now we have it deployed on-prem, but we are planning to move to the cloud very soon. We are using Control-M Workload Archiving, Control-M Enterprise Manager, Control-M servers, agents, APIs, REST APIs, and Control-M Forecast. We use all the services Control-M provides except Control-M Workload Change Manager.

How has it helped my organization?

Since moving to Control-M we have seen a lot of reductions in the manpower needed. For notification, ticketing, and integration, we have different teams. We have Azure teams and some Windows teams. Previously, they were using and managing their own scripts and manually running them. After the migration to Control-M, there were no limitations. Where there are different protocols we can use the APIs and integrate things with Control-M. There are no worries about integrations with Control-M. In UC4 there were lots of limitations because we needed the same protocols to integrate things. With Control-M, there are no such limitations.

In our current environment, there are three sets of applications. The first, an online application, is dependent on some 45 files that have to be generated on Saturday. Our middleware job is supposed to run once all the 45 files have been generated by SAP jobs. There are sequences running through Control-M: First are the SAP jobs that generate the files in a certain location. Once those files are there, the sequence initiates the middleware that moves the files to the proper IT server. All these process flow dependencies go through Control-M very easily.

We have also automated daily audit reports through the solution's reporting facility. Through scripting, we get an alert when anything happens in the Control-M environment. An issue might occur with the agent, the process, or the Control-M server. We have everything reported via email. We can easily see what happened on a given day and sort out any issues.

As a result of using Control-M we have also seen an improvement in Service Level Operations performance. We have some monitoring tools in Control-M and our service SLAs have definitely improved. We have a ticketing system integrated with it and we can easily monitor the SLAs for tickets generated through Control-M. If the person responsible for a ticket will not handle it in the right amount of time, the ticket will pop up with a message saying it's in danger of breaching the SLA. Our service levels are much higher with Control-M, when compared to other tools.

What is most valuable?

Control-M Workload Archiving is a very good feature for us. It helps with our customer requirements in terms of reporting and auditing. We have internal audits every quarter, and every six months we have external audits. During these audits, the auditors get historical data through Control-M. Previously, when we didn't have any archive server, we managed the data in Control-M with man-made scripts, and we would pull the data for the last 365 days, or three or four months back. Since we installed the archiving, we have been able to pull the data, anytime and anywhere, with just one click.

Control-M gives us a unified view where we can easily define, orchestrate, and monitor all our application workflows and data pipelines. We are mostly using SAP and other business warehouse jobs, and we can easily see the systems through Control-M. It gives us a very good view of geographical data. If I go through the Web Services to show things to my customers, they are very satisfied with the Control-M views. They can check historical data and they can see the current view. They can easily pull these up. We are satisfied with the fact that, with one click, we can see all the applications within one view.

Our line-of-business personnel use Control-M’s web interface to support their business initiatives. One of our big applications is our JG application, where a user needs a data pipeline and Power BI jobs with refreshed data. Instead of the user having to send a request to our Control-M team, they can use the Web Services directly to check their data. If they're using an iPad or a desktop, they can easily check on it themselves. They're not dependent on the Control-M team directly. We educate users on how to check things and how to pull the reports. It is very easy to use. Also, we don't have 24/7 support within our company. Suppose a user needs something at midnight. They don't have to wait for the Control-M admin team to give them the report. They can directly pull the details.

What needs improvement?

With the current version update, I'm not sure why we needed a separate database upgrade. Why not put it all in one package? Previously, you could do it either via a manual upgrade or an in-place upgrade but it wasn't separate. But for the current version, we needed to upgrade the database separately. It meant doubling our tasks to do the upgrade. That is something that needs to be improved.

For how long have I used the solution?

I've been using Control-M for the last 17 years. My specialization is in Control-M and I'm very happy and very comfortable with it.

What do I think about the stability of the solution?

I would rate its stability highly, compared to other tools in the market right now, such as UC4, and AutoSys. In the past, I have worked with many banks. All these financial companies are using Control-M, and there is a reason: It's due to the stability.

What do I think about the scalability of the solution?

I would give the scalability a nine out of 10.

In our environment right now, out of 20,000 jobs in Control-M, 15,000 are SAP. We are planning to expand our usage of Control-M to Power BI, Business Warehouse, PeopleSoft, and Azure. Those are in our pipeline right now.

We have about 25,000 users of Control-M on different projects, in the U.S., Japan, India, and Asia Pacific. Some are monitoring programs through Control-M, some are only doing scheduling. Some are responsible for designing, others for the implementation before the licensing. And once this transition team is done, the operations team comes into the picture for monitoring. We have a separate team for integration, as well.

The number of people we require for day-to-day administration of Control-M depends on the job size and the user requirements. We work in an offshore and onsite model. We have a key administrator over the 20,000 jobs, seven schedulers, and nine people on the monitoring team, and that work is done 24/7. The schedulers and admin work 24/5.

How are customer service and technical support?

In a case where we fail to understand an issue by collecting data on our own through our audit reports, we open a case with BMC. BMC always gives us a fast resolution. Their support is very good.

How was the initial setup?

The setup of the current version of Control-M, overall, is very easy. The upgrade is in-place. With one click the agent upgrades, the server upgrades. The only point, as I mentioned, with upgrading, is that we needed a separate database update. When we upgrade our Control-M server, the database server should be upgraded at the same time.

The initial implementation in my current environment was in 2006. When we took over we just upgraded it. After that, we implemented two more Control-M Servers in this environment, as a PoC.

The amount of time required to implement it depends on the environment we are working with. In this environment, we have two production servers, four QA servers, and two testing servers. We have eight Control-M servers, three Control-M Enterprise Manager servers, and more than 400 agents. It depends on the change process. In our change process, we first need to upgrade our QA and test environments. Once that is done, we can go for the production environment the next day. After that, over the next seven days, we update our Control-M agents. Some of the upgrades require downtime. In four to five hours, we could easily update everything, but it's dependent on the downtime and the customer requirements.

When we upgraded to version 20, first we implemented it in our QA environment and we tested the new version in our test environment for three to four months. Once we see there are no bugs, we implement it in our production environment. We've seen a lot of bugs and BMC has had to produce some patches that we have had to apply in our environment. That is why we approach it the way we do in a QA environment, and wait for three months, and then go to production.

When we moved to Control-M, we used the Control-M Conversion Tool. It's a very important tool. It gives us an idea of where we stand. If I'm going to move an old environment to a new environment, it helps us with any errors so that we can rectify them.

What about the implementation team?

Back in 2016, when I was working with version 7, I opened a case with BMC and they helped me to upgrade everything. It was a very good experience. They dedicated a resource to us. We gave them our implementation plan, they reviewed it, and they suggested how to remediate some missing steps. We followed their approach and, at the time of cut-over, they assigned a dedicated resource. If there was an issue, we could open a ticket and they would come online and sort it out. The BMC Assisted MIGration Offering (AMIGO) is a very good program.

What's my experience with pricing, setup cost, and licensing?

You must accept that BMC licensing can be very confusing. No one can easily understand how they calculate things, whether it is user-based, job-based, or server-based. The calculation is quite tough. How BMC calculates licensing is not easily available anywhere. It's a very tough part for the client at times.

But BMC is a market leader, so users don't easily go for different vendors. If there's an option to go with Control-M, they will always choose BMC. But for people who find the licensing challenging, they will go with a different vendor.

For us, the licensing part is managed by a team in the U.S. But what I deal with is that we have to manage our Control-M jobs to a maximum of 30,000, because we have 30,000 licenses. We have 20,000 with fraud detection and 10,000 in non-fraud. There is a BMC utility that can guide you and alert you if the forecast is for an increase beyond the licensing. It will notify us: "Hey, you have a license for 20,000 and the Control-M forecast shows you might need to increase that number in the coming days." So we do some cleanup, some internal housekeeping to remove things and remain under the threshold. Those are some of the things we do as administrators. We try to manage under whatever licensing we have. Through the BMC reporting tool, we can see our peak number of users in a month. BMC charges if you go over a certain peak.

Control-M is very robust. There is no harm to the customer if you choose Control-M every time. But when it comes to licensing, it's very expensive, and sometimes users think twice.

Which other solutions did I evaluate?

Previously we were using UC4 for more than 20,000 jobs. But our customers were not very comfortable with the user interface in UC4. Certain things were not appropriate in that tool. Since our decision to migrate to Control-M, our customers have been very satisfied.

Integration is very easy. When I'm thinking about integrating Control-M with anything I'm not worried about it. I know Control-M will definitely have a way to integrate easily. I have used UC4, AutoSys, and Dollar Universe. But when the requirements include integration, I always think of Control-M, because I know the integration will be very easy. I will never go for any of those other tools.

What other advice do I have?

Control-M is very critical for anyone who is using it.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor. The reviewer's company has a business relationship with this vendor other than being a customer: Partner
PeerSpot user
reviewer1899735 - PeerSpot reviewer
IT - VP at a financial services firm with 10,001+ employees
Real User
We have a better picture of our auditability
Pros and Cons
  • "We have a better picture of our auditability. When someone comes to us, and asks for sources, "How did the deltas occur?" We can provide answers quickly, or at least quicker than what we used to. We are actually sure of the information that we provide, where before it was like, "Hmm, I think it comes from over there. Let me double check, but it gets really convoluted over here and I think that is where it comes from." Now, if it is within the Control-M environment, it has a straightforward answer that we can provide with confidence."
  • "The community and the networking that goes on within that community need improvement. We want to be able to reach out to an SME, and say, "Hey, we are doing it this way. Does that make sense?" Ideally, they come back. and say, "Yes, it does make sense to do it that way. However, if you want to do it this way, then it is a little more efficient." We understand that one solution framework doesn't fit everybody. Depending on the breadth of the data and how broad it is, you may have different models for one over the other."

What is our primary use case?

It is controlling our workflows, ingesting data, and then putting it up into our database platforms. In turn, those are consumed by our internal clients.

We do integrate Control-M Python Client and cloud data service integrations with some of our cloud providers. We have pipelines going out to the public cloud and some pipelines that are internal.

We have public and private cloud channels as well as on-prem. The expectation for most large financial institutions is that we will get 99.9% to the public cloud eventually. We want everything to be in OpEx as opposed to CapEx. We don't want data centers. We just want access to our data and to be able to turn it into information, which in turn, turns it into actionable items. Ideally, we would love to not support any on-prem or hybrid solutions, having everything be public.

How has it helped my organization?

Control-M has improved our visibility and streamlining. We have better clarity into data flows. We can resolve issues faster by not trying to reverse engineer what pipeline the infraction may have come through. We are not completely there yet, but we have better clarity and visibility. 

We have a better picture of our auditability. When someone comes to us, and asks for sources, "How did the deltas occur?" We can provide answers quickly, or at least quicker than what we used to. We are actually sure of the information that we provide, where before it was like, "Hmm, I think it comes from over there. Let me double check, but it gets really convoluted over here and I think that is where it comes from." Now, if it is within the Control-M environment, it has a straightforward answer that we can provide with confidence.

The speed of our audit preparation process is faster. When questions come in about flow, data, or sources, we don't have to try to reverse engineer anything anymore. We are able to go straight to Control-M and find out what the flow is or what happened. The visibility is there. We see the endpoint on this, such as, "What is the reverse flow on it? Where did it come in? Where did that data flow come from?" So, it is not a spaghetti mess anymore. This makes auditability easier. We are able to provide answers more quickly, which in turn, makes the audit process quicker.

Control-M has improved our business service delivery speed. It is more reliable and has increased the release schedules. We are also working on testing standards, and it has shortened the window of getting things to us. It has shortened the window, not to market, but basically getting them live. 

Control-M is critical to our business. If the support ends, we are at risk in some of our critical flows. We have redundancy around it that has been purposely built. We do that with all of our solutions. That way, we are not tied into one specific vendor, then if something happens tomorrow, we don't have a fire drill. We have things in place, but to a certain extent, there is heavy reliance on this solution.

What is most valuable?

The most valuable feature is the Self Service tool. They have metrics in place almost all across the pipeline, which is really nice. 

What needs improvement?

We are not yet really a power user of it. You can take as many training classes as you need, but it is not until you are working with a subject-matter expert (SME) on it that you can find out how you can really make this tool sing. My engineers know how to work Control-M. However, if I ask them, "Oh, is this the most efficient way of doing it?" They may not be able to say, "Yes." It is doing what we want it to do. That is nice and okay, but is it the most efficient, effective way? So, we are not there yet.

For how long have I used the solution?

I have been using it for about four years.

What do I think about the stability of the solution?

The platform is good. We haven't had any major outages. The stability is there.

What do I think about the scalability of the solution?

We really haven't pushed it to any of its limits. No scalability concerns have come up for what we are doing.

If you came to me, saying, "Hey, I was looking at Control-M, but it has some issues." I am going to sit there, and go, "Tell me what the issue is." Right now, we are not using the far reaches of whatever cloud providers are out there. Control-M does well with the major providers.

How are customer service and support?

The community is not as robust as some of our other tools that were replaced. The problem was the other tools that we were using didn't do everything that Control-M is now able to do, like monitoring and the entire pipeline flow.

The community and the networking that goes on within that community need improvement. We want to be able to reach out to an SME, and say, "Hey, we are doing it this way. Does that make sense?" Ideally, they come back. and say, "Yes, it does make sense to do it that way. However, if you want to do it this way, then it is a little more efficient." We understand that one solution framework doesn't fit everybody. Depending on the breadth of the data and how broad it is, you may have different models for one over the other.

I would rate the technical support as seven or eight out of 10.

How would you rate customer service and support?

Positive

Which solution did I use previously and why did I switch?

We had a patchwork set of solutions in place that were getting the job done. The problem with that was we had a lot of SMEs within certain verticals. Therefore, there wasn't one overall picture. Every time we went from one step to another step, we had to start talking to another person to figure out what was going on. So, we were trying to bring everything together under one solution with Control-M.

We are able to have a better picture of our data consumption, e.g., what files or data is brought in. Previously, we would ingest data at different points. The question that would always come back to us would be, "Where did this data come from?" Then, we would always have to reverse engineer and have some documentation on it, but the documentation would be outdated. Someone would change the pipeline and forget to change the documentation. With Control-M, we can see everything in one location. To a certain extent, it is not documentation.

I am an engineer by trade. I have been doing this for over 30 years. I know that it is nice that someone puts together a document describing the environment, but as soon as that document is saved that document is outdated.

We don't throw another tool into the toolbox just because it is a nice pretty tool. We try to figure out what the benefits are. Ideally, in our world, we try to reduce the number of tools because I don't need 50 different screwdrivers in my tool kit. I make sure that I have a flathead and a Phillips, but I don't need 50 screwdrivers. Here, we brought in this solution and it replaced some existing solutions. Now, my engineers don't need to know X number of products. They only need to know half of X number of products.

What about the implementation team?

The tool was vetted by another group before making it available to the organization and putting it into our toolbox. Then, when it was available, we looked to leverage it.

What's my experience with pricing, setup cost, and licensing?

One of the restrictions that we had was with some of the licensing, and not having any insight on the financials part of the product. I don't know what the licensing on the product is, but we don't have an unlimited enterprise license. So, there might be a limitation on either the cost of the licensing or the number of seats.

What other advice do I have?

There is always a learning curve any time you are using a new product. Our engineers who are using Control-M are kind of happy with it. There really are no negatives on its learning curve. I am always weary with new products since it is another thing that someone needs to learn, but now there are other products that we don't use because of Control-M. What I would not be open to is bringing in another product, where we need our engineers to know how to work it and make it efficient as well as support other products already in our environment. So, I like that we can get rid of three or four products and replace them with a single product. As long as the learning curve is not too steep, that is an advantage to me.

We are looking into using Control-M to deliver analytics for complex data. So, the solution is doing either machine learning or complex analytics on top of the data flow. While we do some analytics, it is not to the extent that we really want to.

I would rate this solution as a high seven or low eight (out of 10).

Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
JoseQuintero1 - PeerSpot reviewer
Senior Services Manager at a tech services company with self employed
Real User
We gain speed and reliability because it continuously checks the CRC of the data packages
Pros and Cons
  • "We use Control-M for maintenance on our Oracle and SQL Server databases. It automates maintenance on packages, including standard procedures on the databases themselves, snapshots, checking integrity, verifying the RDBMS of the databases, etc. It ensures they aren't clogged and that they are running smoothly and that there aren't any jobs stuck, eating up the performance of the server or any of the CPU cores."
  • "The performance could be better. Control-M Enterprise Manager tends to slow the system down even on a server with a six-core processor and 32 gigabytes RAM. The console is Java-based, so maybe OpenJDK 16 or 17 would be a performance improvement."

What is our primary use case?

We use Control-M for managed file transfer in the enterprise manager automating database workflows. We're using Oracle Business Intelligence with a generic database like Microsoft SQL Server. Next year, we plan to use Control-M for AWS Natural and upgrade Control-M to Helix Control-M. We need managed file transfers between our servers in Key West and Orlando. Vast amounts of data are routinely transferred between servers for backup and snapshots. 

We will deploy it on AWS and Azure next year. We will also integrate Control-M with other solutions like Salesforce and COS, which is challenging because COS is a complicated legacy IBM OS. Some of our customers require Control-M to do managed file transfer and also COS conversions between IBM jobs. Anything that involves COS or OS/400 is complicated but doable.

How has it helped my organization?

With Managed File Transfer, we gain speed and reliability because a managed transfer continuously checks the CRC of the data packages. That's a God send for those data transfers. Though we are migrating to the cloud, we still use some physical servers for sensitive data from our customers protected with NDAs.

We use Control-M for the maintenance of our Oracle and SQL Server databases. It automates maintenance in packages, including standard procedures on the databases themselves, snapshots, checking integrity, verifying the RDBMS of the databases, etc. It ensures that they aren't clogged, that they run smoothly, and that there aren't any jobs stuck, eating up the performance of the server or any of the CPU cores.

In the past, we had some troubles, and we needed a database admin to keep an eye on it almost 24/7 using the OES. It's essential to ensure everything inside the OES runs smoothly, and there are no stuck jobs or queries eating up table spaces. An admin is still required, but most jobs are now automated. It has had a significant impact on staffing. In the past, we had a couple of DBAs exclusively assigned to Oracle that we were able to reassign to other jobs. 

We reassigned them to other tests and outsourced one to work with our customers. Once we delegated DBA tasks to Control-M for our Oracle databases, we could reassign that DBA as a resource to our client in Puerto Rico. He became a source of income for the company. Also, with time saved by automating all the critical internal business processes, we could dedicate more time and resources to other projects that require human attention. We could devote more resources to projects that advance the company's strategic vision instead of monitoring an Oracle RDBMS 24/7.

If I had to rate how critical Control-M is to our business, I would say it is an eight out of ten. I won't give it a nine because we still rely on older applications, such as Oracle databases, but an orchestrator will always get at least an eight on our book. For speed of process execution, I would say it is a nine out of ten. Previously, it was a four, and now it's a nine.

What is most valuable?

Control-M is intuitive, and BMC has tutorials for every application to help you with the basics. Once you know what you're doing, everything falls into place. The graphical interface is drag and drop. There are plenty of objects to drag and drop inside. You need to study them, but once you know how it works, it's just dragging and dropping like you are playing with a Lego set. 

You drag two actions to the workspace and connect them to establish a relationship, schedules, and subtasks inside each one of them. It seems complex initially, but it becomes intuitive the longer you use it.

You can almost reach out intuitively into every nook and cranny of the entire UI. It's user-friendly for the initiate, but you could be lost if you've never used an orchestrator or an enterprise-grade software like Control-M. However, Control-M has built-in tutorials that help you with the first steps. The tutorial isn't comprehensive, but at least you will learn the first steps, so you can advance and learn more.

What needs improvement?

The performance could be better. Control-M Enterprise Manager tends to slow the system down even on a server with a six-core processor and 32 gigabytes of RAM. The console is Java-based, so maybe OpenJDK 16 or 17 would be a performance improvement.

For how long have I used the solution?

I've been using Control-M for five years.

What do I think about the scalability of the solution?

The scalability is quite good. The framework lets you start with Control-M Enterprise Manager and add other products as you see fit. We added MFT, then Control-M for databases, and Oracle Business Intelligence. One of our customers added Control-M, including the agent for IBMI and another for Azure.

How are customer service and support?

I rate BMC support a solid nine out of ten. I say nine because I never give a ten to anyone. 

How would you rate customer service and support?

Positive

Which solution did I use previously and why did I switch?

Before Control-M, we had a traditional solution using an FTP server. Even with a T1 line that provided almost gigabyte speed, we still had artifacts during the transfer that corrupted the data. It caused serious problems when transferring 30 gigabytes of a necessary backup overnight that failed on gigabyte 28 because it was corrupted. Still, the mirror server rejected the mage because of corruption when we tried to restore it. 

We had to rely on traditional monitoring tools like SolarWinds and IBM solutions, which are pretty expensive. These tools only monitor, so they're typically not reactive or able to orchestrate the steps of a workflow. They don't follow up on each step inside the workflow, notify you when a step completes, or send alerts when something gets stuck and requires action.

How was the initial setup?

I'm the senior services manager, and overseeing the deployment of Control-M is part of my job. I did not install it, but I supervised the team. It was straightforward because we all got our BMC certification before the deployment. Our team included me and two technicians. We also had a DBA around to integrate the database.

What about the implementation team?

We did everything ourselves with some occasional help from BMC support. We emailed them a couple of times to check something, but so far, everything has gone smoothly. 

What was our ROI?

We recovered our initial investment in six months and were ready to commit more, so we could recover more. We saw an ROI with Control-M in the first two years because we could take a DBA off monitoring databases and loan them out to another company while saving time by speeding up these processes. 

Control-M gives us a lot of flexibility to automate our time-sensitive and data-critical processes. This is essential for enterprises, but Control-M isn't for everyone. Small and medium-sized businesses can use Control-M, but its power can only be leveraged by large enterprises because of the complexity of their business processes and the sheer size of data they handle. I think enterprise companies are the only ones that see an effective ROI from purchasing a tool like BMC Control-M for automating their business processes.

Small companies that purchase Control-M often cancel the contract after six months because it's too expensive, and they can get the job done using traditional methods. That's okay. It's about the business processes' complexity, depth, and maturity. 

What's my experience with pricing, setup cost, and licensing?

The price is right because of the licensing schema, which is based on nodes and processes. You purchase what you use, no more and no less, and you can grow with time. 

Which other solutions did I evaluate?

There are few options like Control-M in the market, and the closest competitors are far more expensive.

What other advice do I have?

I rate Control-M a nine out of ten. Control-M is flexible. You can use it in Azure, and they have a generic option for the cloud. You can deploy it in your own private cloud or on other cloud solutions like Kubernetes. You can use Control-M for big data applications like IBM InfoSphere. There's a Control-M solution for almost any situation.

There is so much to learn on the backend of the business processes. Typically when you see a business process, you only see a workflow, like a flow chart, arrows, boxes, etc. However, there's a whole new world under the hood. It's crucial to dig deeper and learn how to improve the processes. It's like you become the mechanic of your own car. The more you understand the engine, the more you can tweak it to get more speed, gas mileage, performance, strength, horsepower, etc. Control-M almost compels you to learn about that.

It's user-friendly, but you need some training. We have a certification from BMC. You need some prior training specifically in Control-M Enterprise Manager to know what you're doing because it's delicate. There are so many ways to customize job creation, automation, monitoring, etc. that you need at least a crash course on creating a job, monitoring, setting up alarms, and building workflows. 

It should take you no more than a week to get the hang of it, and there's BMC University, where you can get free training to use Control-M. Once you know the basics, Control-M practically handles itself. It's intuitive once you figure it out. 

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
Buyer's Guide
Download our free Control-M Report and get advice and tips from experienced pros sharing their opinions.
Updated: June 2025
Buyer's Guide
Download our free Control-M Report and get advice and tips from experienced pros sharing their opinions.