After switching a personnal solution from .NET 6 to .NET 7, time to read large amount of data go from 18s to 4min 30s (approximately).
Before starting to develop it, I have a test function allowing me to have the critical path without big bottlenecks.
private void SpeedTest()
{
int nbdata = 6000000;
List<int> list = new(nbdata);
var rnd = RandomNumberGenerator.Create();
Random rand = new(12345);
for (int i = 0; i < nbdata; i++)
{
var rnddata = new byte[sizeof(int)];
rnd.GetBytes(rnddata);
list.Add(BitConverter.ToInt32(rnddata));
}
int[] arr = list.ToArray();
//Begin test
int chk = 0;
Stopwatch watch = Stopwatch.StartNew();
for (int rpt = 0; rpt < 100; rpt++)
{
int len = list.Count;
for (int i = 0; i < len; i++)
{
chk += list[i];
}
}
watch.Stop();
SpeedText.Text += string.Format("List/for Count out: {0}ms ({1})", watch.ElapsedMilliseconds, chk) + Environment.NewLine;
chk = 0;
watch = Stopwatch.StartNew();
for (int rpt = 0; rpt < 100; rpt++)
{
for (int i = 0; i < list.Count; i++)
{
chk += list[i];
}
}
watch.Stop();
SpeedText.Text += string.Format("List/for Count in: {0}ms ({1})", watch.ElapsedMilliseconds, chk) + Environment.NewLine;
chk = 0;
watch = Stopwatch.StartNew();
for (int rpt = 0; rpt < 100; rpt++)
{
int len = arr.Length;
for (int i = 0; i < len; i++)
{
chk += arr[i];
}
}
watch.Stop();
SpeedText.Text += string.Format("Array/for Count out: {0}ms ({1})", watch.ElapsedMilliseconds, chk) + Environment.NewLine;
chk = 0;
watch = Stopwatch.StartNew();
for (int rpt = 0; rpt < 100; rpt++)
{
for (int i = 0; i < arr.Length; i++)
{
chk += arr[i];
}
}
watch.Stop();
SpeedText.Text += string.Format("Array/for Count in: {0}ms ({1})", watch.ElapsedMilliseconds, chk) + Environment.NewLine;
chk = 0;
watch = Stopwatch.StartNew();
for (int rpt = 0; rpt < 100; rpt++)
{
int k = list.Count;
for (int j = 0; j < k; j++)
{
chk += list[j];
}
}
watch.Stop();
SpeedText.Text += string.Format("List/for: {0}ms ({1})", watch.ElapsedMilliseconds, chk) + Environment.NewLine;
chk = 0;
watch = Stopwatch.StartNew();
for (int rpt = 0; rpt < 100; rpt++)
{
foreach (int i in list)
{
chk += i;
}
}
watch.Stop();
SpeedText.Text += string.Format("List/foreach: {0}ms ({1})", watch.ElapsedMilliseconds, chk) + Environment.NewLine;
chk = 0;
watch = Stopwatch.StartNew();
for (int rpt = 0; rpt < 100; rpt++)
{
list.ForEach(i => chk += i);
}
watch.Stop();
SpeedText.Text += string.Format("List/foreach function: {0}ms ({1})", watch.ElapsedMilliseconds, chk) + Environment.NewLine;
chk = 0;
watch = Stopwatch.StartNew();
for (int rpt = 0; rpt < 100; rpt++)
{
int k = arr.Length;
for (int j = 0; j < k; j++)
{
chk += arr[j];
}
}
watch.Stop();
SpeedText.Text += string.Format("Array/for: {0}ms ({1})", watch.ElapsedMilliseconds, chk) + Environment.NewLine;
chk = 0;
watch = Stopwatch.StartNew();
for (int rpt = 0; rpt < 100; rpt++)
{
foreach (int i in arr)
{
chk += i;
}
}
watch.Stop();
SpeedText.Text += string.Format("Array/foreach: {0}ms ({1})", watch.ElapsedMilliseconds, chk) + Environment.NewLine;
}
Result .NET 6:
List/for Count out: 1442ms (398007896)
List/for Count in: 1446ms (398007896)
Array/for Count out: 1256ms (398007896)
Array/for Count in: 1254ms (398007896)
List/for: 1435ms (398007896)
List/foreach: 1258ms (398007896)
List/foreach function: 1452ms (398007896) <=
Array/for: 1255ms (398007896)
Array/foreach: 1254ms (398007896)
Result .NET 7:
List/for Count out: 1483ms (272044760)
List/for Count in: 1489ms (272044760)
Array/for Count out: 1255ms (272044760)
Array/for Count in: 1263ms (272044760)
List/for: 1482ms (272044760)
List/foreach: 1873ms (272044760)
List/foreach function: 7997ms (272044760) <=
Array/for: 1254ms (272044760)
Array/foreach: 1255ms (272044760)
Code behind this issue:
list.ForEach(i => chk += i);
Is this problem inside .NET 7?
Do I have any hope of finding solution without having to change all calls to this function?
I use many other functions that perform better in .NET 7 than in .NET 6.
I wish I could stay on this version.
What do you recommend?
Thanks.
EDIT:
I used ForEach few times to reading code. Initially in .NET 6, the time loss was acceptable. I used Tuple with data read in large files.
Example:
listValue.ForEach(x => process((new col(x.name, position++, startId++, x.refState, x.refPosition, x.refTable, x.withoutRef, x.deleted, x.resetData), option)));
foreach((string name, uint refState, uint refPosition, uint refTable, bool withoutRef, bool deleted, bool resetData)x in listValue)
{
process((new col(x.name, position++, startId++, x.refState, x.refPosition, x.refTable, x.withoutRef, x.deleted, x.resetData), option))
};
My software is far from being finished and I use public data files to test it:
- xlsx file with 1,000,000 rows by 14 columns
- csv file with 10,000,000 rows by 14 columns
I made some changes in my code between switching to .NET6 and .NET 7 and saw the times increase dramatically on my 1st test in .NET 7.
So I went back to my original benchmark code to see if there were any changes before revising my entire code.
I think processing time and code of this benchmark is suitable to make a decision in my case. I just want to see how long the user will have to wait. So I put myself in the same case as user. Benchmark with List of 5000 elements is not relevant. I work with large List and performance can be affected by this size.
This test is basic and reveals a big difference between .NET 6 and .NET 7 with same code.
Performance of Arrays vs. Lists
Here issue is not how measurement is made but the result.
There is no library used that could have differents versions and that could influence the result.
I test on Windows 10 with Ryzen 1700 and RAM 16Gb.
EDIT2:
Project to test:
https://github.com/gandf/TestPerfForEach
Clean & generate project and run outside Visual Studio.
Result .NET 6:
Test with 6000000 NbData
List/foreach: 1254ms (2107749308)
List/foreach function: 1295ms (2107749308)
Test with 6000000 NbData
List/foreach: 1259ms (1107007452)
List/foreach function: 1255ms (1107007452)
Test with 6000000 NbData
List/foreach: 1253ms (745733412)
List/foreach function: 1256ms (745733412)
Test with 6000000 NbData
List/foreach: 1253ms (-280872836)
List/foreach function: 1259ms (-280872836)
Result .NET 7:
Test with 6000000 NbData
List/foreach: 1866ms (-998431744)
List/foreach function: 8347ms (-998431744)
Test with 6000000 NbData
List/foreach: 1753ms (715062008)
List/foreach function: 1368ms (715062008)
Test with 6000000 NbData
List/foreach: 1754ms (667927108)
List/foreach function: 1335ms (667927108)
Test with 6000000 NbData
List/foreach: 1749ms (310491380)
List/foreach function: 1366ms (310491380)
Same condition and tests run several times:
- .NET 6 faster.
- Issue with
list.ForEach
only first run. After quicker thanforeach
.
2
Answers
I found source of this issue. One year ago, I see this : https://devblogs.microsoft.com/dotnet/performance-improvements-in-net-6/
I left in PATH 2 options:
With these parameters, I had noticed a very slight degradation of the first call in .NET 6 with an improvement with others calls. So I kept it because the impact was negligible.
So there is a case where the first call takes x6 more time in .NET7 with these options.
I just deleted them. Results after reboot:
.NET 6
.NET 7
It's fixed.
Using
BenchmarkDotNet
, I tried re-creating your scenario and then ran it against both .NET6 and .NET7.I used smaller numbers because the benchmarking tool can take a minute.
Here’s the code I used:
Here is the
Program.cs
on my console app:Here are my results:
.NET 6
.NET 7
To your point, the List
ForEach
does appear to have slowed down a bit between the two versions.Those numbers are in NANOseconds, so the change is pretty small (~50 ns).
The other numbers all seem to have improved between versions. Memory allocation held steady.