Using the "Split CSV evenly between Test engines" option in Azure Load Tests corrupts the data

Scalise, Michael 20 Reputation points
2024-10-28T18:58:59.5366667+00:00

I have an Azure load test that takes a CSV file with a set of simulated devices and their test credentials to authenticate in a QA environment. Using this test file I was able to scale up to 1,000 test users, but to add more users than that I needed to increase the number of test engines and split the CSV file among them. There are more than enough rows in the CSV to accommodate the number of users that I am trying to replicate, and the files does not have a header row. These appear to be the only troubleshooting suggestions in the existing documentation.

However, when I split the file I start getting corrupted data back. Instead of seeing fields populated with "TestDevice1" I see long, gabled like "䉁穁佂䅩硳䡲歈桁牔杉". These gibberish strings are also extremely long, far longer than the fields in the original CSV.

To make matters worse, I can't use debug mode because it limits the Load Test to one test engine, nor can I see what test files are actually being passed to the engines because if I click "Download Input Files" after the test completes then it gives me back the original, pre-split versions of the input files.

Does anyone know what causes the Split CSV option to corrupt files like this?

Azure Load Testing
Azure Load Testing
An Azure service that enables developers and testers to generate insights on how to improve the performance, scalability, and capacity usage of their application
68 questions
0 comments No comments
{count} votes

Accepted answer
  1. RevelinoB 3,435 Reputation points
    2024-10-28T20:25:50.7566667+00:00

    Hi Michael,

    I've seen this issue before with Azure Load Testing, and it can be a bit tricky. The corrupted data—like the long garbled strings you're seeing—often stems from encoding or how Azure handles the CSV split. Here are some approaches that usually help:

     

    Encoding Check: Sometimes, the problem comes from the file’s encoding. Azure Load Testing works best with CSVs saved as UTF-8 or UTF-8 BOM. If you haven’t already, try re-saving your CSV in UTF-8 and re-upload it. Tools like Notepad++ can quickly confirm and change encoding if needed.

    Header Row Experiment: Even though your file doesn’t need a header, adding a dummy header row sometimes helps the splitter allocate rows properly. It's worth a quick test to see if it resolves the garbling.

    Manual Splitting: To bypass Azure’s CSV splitting (which seems to be the root issue here), you could manually divide the CSV into smaller files—say, 500 rows each—and upload each file to a separate engine. This approach often avoids the garbling altogether.

    CSV Format Consistency: Double-check the CSV for any irregularities, like inconsistent delimiters or stray quotes. Even small formatting hiccups can throw off the splitting process and cause these odd strings.

    Consider External Data Source: If Azure Load Testing supports pulling data from an external source, using something like Azure Storage could give you more control, especially as you scale up. This way, you wouldn’t need to worry about CSV splitting at all.

     

    If none of these work, reaching out to Azure Support is a good call. They might have insights into any quirks or recent issues with the CSV splitter. Let me know if any of these solutions help, or if you'd like to dive deeper into one!

    1 person found this answer helpful.
    0 comments No comments

0 additional answers

Sort by: Most helpful

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.