What is our primary use case?
We conduct economic research focused on Israel's tech industry and innovation policy. Our work mainly concerns understanding the Israeli tech sector. We're using data under the BYT initiative, but I am not actively using any scrapers. I only download data already collected and generously provided to us for free.
I use two datasets: LinkedIn profiles for individuals and LinkedIn profiles for companies. While we are exploring other datasets, these are the ones we've been using so far. We analyze this data to support various research projects, such as understanding the characteristics of employees in the tech industry. For example, we investigate how many employees companies are adding or losing, what tech workers study, and what kind of education is most likely to lead to a career in tech, including within different industry subsectors.
What is most valuable?
There are a lot more advanced options. It looks very impressive.
What needs improvement?
It would be better if variables like experience or education containing complex data were already organized into sub-variables. For example, the experience variable could be broken down into company name, company ID, job description, start year, and end year.
As part of the Bright Initiative, since we’re a nonprofit, we are limited in the number of data points we can download. This can occasionally be challenging, but we manage to work around it. In the past, I encountered a few other challenges as well. Initially, I had difficulty transferring the data from Bright Data to our cloud.
One past issue involved date formats. There were many different date formats for the same data points. LinkedIn adjusts the date format based on the user’s browsing location. So, if you're viewing a profile from the US, the date might appear in a month-day-year format, whereas in other regions, it could show up as day-month-year.
When using automatic web crawlers to download data, the data formats can vary depending on the server's IP address running the scraper, not your location.
I received the employment start date variable in about fifty different formats, which I had to standardize. I wrote some code to convert all the time formats into a consistent format.
For how long have I used the solution?
I have been using Bright Data for two years.
Buyer's Guide
Bright Data
March 2026
Learn what your peers think about Bright Data. Get advice and tips from experienced pros sharing their opinions. Updated: March 2026.
884,797 professionals have used our research since 2012.
What do I think about the stability of the solution?
There were no issues with stability.
What do I think about the scalability of the solution?
We have two researchers working on this and some research assistants.
How are customer service and support?
We asked for help understanding some of the data on a few occasions. They were very professional and accommodating. Even though we’re not a paid customer, they genuinely went out of their way to help solve our problems. In a few cases, we asked if they could create new features for us, and they were willing to consider it, which was very nice. Overall, it was an excellent experience, and they quickly replied.
How was the initial setup?
The filters used to retrieve and download the data are quite straightforward. It probably takes about an hour or two to go through the system and fully understand how it works. You can figure it out; it's simple.
For anyone working with data professionally, it's a manageable process. It requires some time investment, but it's not difficult, and someone experienced in data management should be able to get comfortable with it quickly.
LinkedIn profile data is organized in JSON file format. Understanding the structure of this data can be a bit tricky. For example, there's a variable called experience. Under this variable, a text string contains all the information the user has entered about their professional experience, similar to a resume. For some reason, possibly due to LinkedIn changing its format at some point, there are two different formats used for that experience string. To handle this, I wrote a small code that parses the string correctly and creates a table with all the necessary information. However, this took a lot of work—it involved thoroughly understanding the data, writing the code, and testing it to ensure it worked as intended. Some variables, like the country code, are straightforward. Other variables contain text strings that are difficult to decode from the JSON format into something useful. Extracting that information requires more effort.
What's my experience with pricing, setup cost, and licensing?
We are using the free version. It’s expensive. We’re a small nonprofit with limited resources. Our work is focused on research for the public good, so we don’t have a lot of budget to spare.
What other advice do I have?
We hired another company to scrape the data for us. The quality of the data we received was similar to what we had before, but accessing the data without doing it ourselves has been much more useful. When handling it independently, paying someone was much more expensive, and we were limited to datasets of only a few hundred thousand individuals.
We're working with datasets that include millions of profiles. For example, the number of Israeli LinkedIn profiles is about 1.8 million, and that's the scale of data we’re using now. I like the interface and the ability to apply filters to create specific subsets according to my research needs at any time. Downloading the data is straightforward, with many transfer options to various cloud services.
The support is very professional and excellent. If I need a data scrape, I’d consider them the first place to go.
Overall, I rate the solution a ten out of ten.
Disclosure: My company does not have a business relationship with this vendor other than being a customer.