@@ -759,29 +759,17 @@ let ParseInputFilesInParallel (tcConfig: TcConfig, lexResourceManager, sourceFil
759759 for fileName in sourceFiles do
760760 checkInputFile tcConfig fileName
761761
762- // Order files to be parsed by size (descending). The idea is to process big files first,
763- // so that near the end when only some nodes are still processing items, it's the smallest items,
764- // which should reduce the period of time where only some nodes are busy.
765- // This requires some empirical evidence.
766- let sourceFiles =
767- sourceFiles
768- |> List.mapi ( fun i f -> i, f)
769- |> List.sortBy ( fun ( _i , f ) -> - FileInfo( f) .Length)
770-
771762 let sourceFiles = List.zip sourceFiles isLastCompiland
772763
773764 UseMultipleDiagnosticLoggers ( sourceFiles, delayLogger, None) ( fun sourceFilesWithDelayLoggers ->
774765 sourceFilesWithDelayLoggers
775- |> ListParallel.map ( fun ((( idx , fileName ) , isLastCompiland ), delayLogger ) ->
766+ |> ListParallel.map ( fun (( fileName , isLastCompiland ), delayLogger ) ->
776767 let directoryName = Path.GetDirectoryName fileName
777768
778769 let input =
779770 parseInputFileAux ( tcConfig, lexResourceManager, fileName, ( isLastCompiland, isExe), delayLogger, retryLocked)
780771
781- idx, ( input, directoryName))
782- // Bring back index-based order
783- |> List.sortBy fst
784- |> List.map snd)
772+ ( input, directoryName)))
785773
786774let ParseInputFilesSequential ( tcConfig : TcConfig , lexResourceManager , sourceFiles , diagnosticsLogger : DiagnosticsLogger , retryLocked ) =
787775 let isLastCompiland , isExe = sourceFiles |> tcConfig.ComputeCanContainEntryPoint
0 commit comments