From 58d2dd0cdffeb4a49184bbfcf7479c1172eac825 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Thu, 21 Mar 2024 09:39:36 -0700 Subject: [PATCH 01/26] Make tests and product target net8 instead of net6 Net6 & net7 will be out of support by the time we release. --- .vsts-dotnet-ci.yml | 16 ++++++++-------- build/.night-build.yml | 4 ++-- build/Codecoverage.proj | 2 +- build/ci/job-template.yml | 4 ++-- build/codecoverage-ci.yml | 2 +- .../Microsoft.ML.Samples.GPU.csproj | 2 +- .../Microsoft.Data.Analysis.Interactive.csproj | 2 +- .../Microsoft.Data.Analysis.csproj | 4 ++-- .../Microsoft.ML.AutoML.Interactive.csproj | 2 +- .../Microsoft.ML.AutoML.csproj | 4 ++-- .../Templates/Console/PredictProject.cs | 2 +- .../Templates/Console/PredictProject.tt | 2 +- .../Microsoft.ML.Console.csproj | 2 +- .../Microsoft.ML.CpuMath.csproj | 4 ++-- .../Microsoft.ML.DnnAnalyzer.csproj | 2 +- src/Native/Native.proj | 2 +- test/Directory.Build.props | 5 ++--- ...rosoft.Data.Analysis.Interactive.Tests.csproj | 2 +- ...crosoft.Data.Analysis.PerformanceTests.csproj | 2 +- .../Microsoft.ML.CpuMath.PerformanceTests.csproj | 8 +------- .../Microsoft.ML.CpuMath.UnitTests.csproj | 2 +- 21 files changed, 34 insertions(+), 41 deletions(-) diff --git a/.vsts-dotnet-ci.yml b/.vsts-dotnet-ci.yml index 90f561acbd..a9a62df728 100644 --- a/.vsts-dotnet-ci.yml +++ b/.vsts-dotnet-ci.yml @@ -51,12 +51,12 @@ jobs: _configuration: Debug _config_short: DI _includeBenchmarkData: false - _targetFramework: net6.0 + _targetFramework: net8.0 Release_Build: _configuration: Release _config_short: RI _includeBenchmarkData: true - _targetFramework: net6.0 + _targetFramework: net8.0 innerLoop: true pool: name: NetCore-Public @@ -74,12 +74,12 @@ jobs: _configuration: Debug _config_short: DI _includeBenchmarkData: false - _targetFramework: net6.0 + _targetFramework: net8.0 Release_Build: _configuration: Release _config_short: RI _includeBenchmarkData: true - _targetFramework: net6.0 + _targetFramework: net8.0 innerLoop: true pool: name: NetCore-Public @@ -127,12 +127,12 @@ jobs: _configuration: Debug _config_short: DI _includeBenchmarkData: false - _targetFramework: net6.0 + _targetFramework: net8.0 Release_Build: _configuration: Release _config_short: RI _includeBenchmarkData: true - _targetFramework: net6.0 + _targetFramework: net8.0 innerLoop: true pool: vmImage: macOS-12 @@ -148,12 +148,12 @@ jobs: _configuration: Debug _config_short: DI _includeBenchmarkData: false - _targetFramework: net6.0 + _targetFramework: net8.0 Release_Build: _configuration: Release _config_short: RI _includeBenchmarkData: false - _targetFramework: net6.0 + _targetFramework: net8.0 innerLoop: true vsTestConfiguration: "/Framework:.NETCoreApp,Version=v6.0" pool: diff --git a/build/.night-build.yml b/build/.night-build.yml index d17393bc46..6f2f8cf512 100644 --- a/build/.night-build.yml +++ b/build/.night-build.yml @@ -37,12 +37,12 @@ jobs: Debug_Build: _configuration: Debug _config_short: DI - _targetFramework: net6.0 + _targetFramework: net8.0 _includeBenchmarkData: false Release_Build: _configuration: Release _config_short: RI - _targetFramework: net6.0 + _targetFramework: net8.0 _includeBenchmarkData: true nightlyBuild: true pool: diff --git a/build/Codecoverage.proj b/build/Codecoverage.proj index 495323f85e..39183a168b 100644 --- a/build/Codecoverage.proj +++ b/build/Codecoverage.proj @@ -2,7 +2,7 @@ - net6.0 + $(NetMinimum) diff --git a/build/ci/job-template.yml b/build/ci/job-template.yml index 4c033c30cd..5b46ca04ae 100644 --- a/build/ci/job-template.yml +++ b/build/ci/job-template.yml @@ -51,12 +51,12 @@ jobs: _configuration: Debug _config_short: D _includeBenchmarkData: false - _targetFramework: net6.0 + _targetFramework: net8.0 Release_Build: _configuration: Release _config_short: R _includeBenchmarkData: true - _targetFramework: net6.0 + _targetFramework: net8.0 ${{ if ne(parameters.customMatrixes, '') }}: ${{ insert }}: ${{ parameters.customMatrixes }} diff --git a/build/codecoverage-ci.yml b/build/codecoverage-ci.yml index 96f3a02cff..2f610b6c0e 100644 --- a/build/codecoverage-ci.yml +++ b/build/codecoverage-ci.yml @@ -35,7 +35,7 @@ jobs: _configuration: Debug _config_short: D _includeBenchmarkData: false - _targetFramework: net6.0 + _targetFramework: net8.0 codeCoverage: true pool: name: NetCore-Public diff --git a/docs/samples/Microsoft.ML.Samples.GPU/Microsoft.ML.Samples.GPU.csproj b/docs/samples/Microsoft.ML.Samples.GPU/Microsoft.ML.Samples.GPU.csproj index 4c1cb229e4..0483f0bfbf 100644 --- a/docs/samples/Microsoft.ML.Samples.GPU/Microsoft.ML.Samples.GPU.csproj +++ b/docs/samples/Microsoft.ML.Samples.GPU/Microsoft.ML.Samples.GPU.csproj @@ -1,7 +1,7 @@  - net6.0 + net8.0 Exe false diff --git a/src/Microsoft.Data.Analysis.Interactive/Microsoft.Data.Analysis.Interactive.csproj b/src/Microsoft.Data.Analysis.Interactive/Microsoft.Data.Analysis.Interactive.csproj index 3a5b1aaf60..939fffd817 100644 --- a/src/Microsoft.Data.Analysis.Interactive/Microsoft.Data.Analysis.Interactive.csproj +++ b/src/Microsoft.Data.Analysis.Interactive/Microsoft.Data.Analysis.Interactive.csproj @@ -1,7 +1,7 @@  - net6.0 + $(NetMinimum) false diff --git a/src/Microsoft.Data.Analysis/Microsoft.Data.Analysis.csproj b/src/Microsoft.Data.Analysis/Microsoft.Data.Analysis.csproj index 1badcdef5c..23f7d079c5 100644 --- a/src/Microsoft.Data.Analysis/Microsoft.Data.Analysis.csproj +++ b/src/Microsoft.Data.Analysis/Microsoft.Data.Analysis.csproj @@ -21,8 +21,8 @@ We cannot include it with a ProjectReference, since that would create a circular dependency. If M.D.A.I undergoes TFM or dependency changes, we need to update the TargetFramework passed in below. --> - - + + diff --git a/src/Microsoft.ML.AutoML.Interactive/Microsoft.ML.AutoML.Interactive.csproj b/src/Microsoft.ML.AutoML.Interactive/Microsoft.ML.AutoML.Interactive.csproj index c391b0a00b..3c067911a5 100644 --- a/src/Microsoft.ML.AutoML.Interactive/Microsoft.ML.AutoML.Interactive.csproj +++ b/src/Microsoft.ML.AutoML.Interactive/Microsoft.ML.AutoML.Interactive.csproj @@ -1,7 +1,7 @@  - net6.0 + $(NetMinimum) false $(NoWarn) diff --git a/src/Microsoft.ML.AutoML/Microsoft.ML.AutoML.csproj b/src/Microsoft.ML.AutoML/Microsoft.ML.AutoML.csproj index 909139b33b..b101149b95 100644 --- a/src/Microsoft.ML.AutoML/Microsoft.ML.AutoML.csproj +++ b/src/Microsoft.ML.AutoML/Microsoft.ML.AutoML.csproj @@ -17,8 +17,8 @@ - - + + diff --git a/src/Microsoft.ML.CodeGenerator/Templates/Console/PredictProject.cs b/src/Microsoft.ML.CodeGenerator/Templates/Console/PredictProject.cs index 8acec761a7..86fca91345 100644 --- a/src/Microsoft.ML.CodeGenerator/Templates/Console/PredictProject.cs +++ b/src/Microsoft.ML.CodeGenerator/Templates/Console/PredictProject.cs @@ -27,7 +27,7 @@ internal partial class PredictProject : PredictProjectBase public virtual string TransformText() { this.Write("\r\n\r\n \r\n Exe\r\n net6.0\r\n \r\n net8.0\r\n \r\n \r\n \r\n"); diff --git a/src/Microsoft.ML.CodeGenerator/Templates/Console/PredictProject.tt b/src/Microsoft.ML.CodeGenerator/Templates/Console/PredictProject.tt index a23b575ccf..b12ca819be 100644 --- a/src/Microsoft.ML.CodeGenerator/Templates/Console/PredictProject.tt +++ b/src/Microsoft.ML.CodeGenerator/Templates/Console/PredictProject.tt @@ -8,7 +8,7 @@ Exe - net6.0 + net8.0 diff --git a/src/Microsoft.ML.Console/Microsoft.ML.Console.csproj b/src/Microsoft.ML.Console/Microsoft.ML.Console.csproj index 78a1e768af..3908f25f19 100644 --- a/src/Microsoft.ML.Console/Microsoft.ML.Console.csproj +++ b/src/Microsoft.ML.Console/Microsoft.ML.Console.csproj @@ -1,7 +1,7 @@  - net6.0 + $(NetMinimum) Exe MML Microsoft.ML.Tools.Console.Console diff --git a/src/Microsoft.ML.CpuMath/Microsoft.ML.CpuMath.csproj b/src/Microsoft.ML.CpuMath/Microsoft.ML.CpuMath.csproj index f6f7abb7da..83b4bffc64 100644 --- a/src/Microsoft.ML.CpuMath/Microsoft.ML.CpuMath.csproj +++ b/src/Microsoft.ML.CpuMath/Microsoft.ML.CpuMath.csproj @@ -1,7 +1,7 @@  - netstandard2.0;net6.0 + netstandard2.0;$(NetMinimum) true Microsoft.ML.CpuMath contains optimized math routines for ML.NET. true @@ -14,7 +14,7 @@ - + diff --git a/src/Microsoft.ML.DnnAnalyzer/Microsoft.ML.DnnAnalyzer/Microsoft.ML.DnnAnalyzer.csproj b/src/Microsoft.ML.DnnAnalyzer/Microsoft.ML.DnnAnalyzer/Microsoft.ML.DnnAnalyzer.csproj index 252cce0312..af01c14dff 100644 --- a/src/Microsoft.ML.DnnAnalyzer/Microsoft.ML.DnnAnalyzer/Microsoft.ML.DnnAnalyzer.csproj +++ b/src/Microsoft.ML.DnnAnalyzer/Microsoft.ML.DnnAnalyzer/Microsoft.ML.DnnAnalyzer.csproj @@ -2,7 +2,7 @@ Exe - net6.0 + $(NetMinimum) DnnAnalyzer diff --git a/src/Native/Native.proj b/src/Native/Native.proj index 25892958cf..ba075571bd 100644 --- a/src/Native/Native.proj +++ b/src/Native/Native.proj @@ -185,7 +185,7 @@ Include="$(NativeAssetsBuiltPath)\$(NativeLibPrefix)CpuMathNative$(NativeLibExtension)" RelativePath="Microsoft.ML.CpuMath\runtimes\$(PackageRid)\nativeassets\netstandard2.0" /> + RelativePath="Microsoft.ML.CpuMath\runtimes\$(PackageRid)\nativeassets\$(NetMinimum)" /> diff --git a/test/Directory.Build.props b/test/Directory.Build.props index 51d5e2eb15..980a2dbc35 100644 --- a/test/Directory.Build.props +++ b/test/Directory.Build.props @@ -2,9 +2,8 @@ - net6.0;net462;net8.0 - net6.0 - net6.0 + $(NetMinimum);$(NetCurrent) + $(TargetFrameworks);$(NetFrameworkMinimum) $(TestTargetFramework) win-x64 true diff --git a/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj b/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj index c49ffd82d1..44f6d29bc7 100644 --- a/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj +++ b/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj @@ -1,7 +1,7 @@  - net6.0 + $(NetMinimum);$(NetCurrent) $(NoWarn);MSML_ExtendBaseTestClass diff --git a/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj b/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj index fa3a7b571a..39e6f38efa 100644 --- a/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj +++ b/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj @@ -3,7 +3,7 @@ Exe disable - net6.0 + $(NetMinimum);$(NetCurrent) false diff --git a/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj b/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj index 525256afff..802c72c729 100644 --- a/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj +++ b/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj @@ -3,13 +3,7 @@ Exe false - net6.0 - - + $(NetMinimum);$(NetCurrent) true false true diff --git a/test/Microsoft.ML.CpuMath.UnitTests/Microsoft.ML.CpuMath.UnitTests.csproj b/test/Microsoft.ML.CpuMath.UnitTests/Microsoft.ML.CpuMath.UnitTests.csproj index a1fc06863d..0a4895777b 100644 --- a/test/Microsoft.ML.CpuMath.UnitTests/Microsoft.ML.CpuMath.UnitTests.csproj +++ b/test/Microsoft.ML.CpuMath.UnitTests/Microsoft.ML.CpuMath.UnitTests.csproj @@ -5,7 +5,7 @@ - + From 2e7f7680b3d86e268d415d463a4d99ee8c9b308f Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Thu, 21 Mar 2024 09:40:01 -0700 Subject: [PATCH 02/26] Remove private copy of RemoteExecutor --- Directory.Build.props | 2 +- Microsoft.ML.sln | 11 - .../Microsoft.ML.AutoML.Samples.csproj | 2 +- .../Microsoft.ML.Samples.csproj | 2 +- eng/Version.Details.xml | 3 +- eng/Versions.props | 2 +- eng/helix.proj | 2 +- ...CodeGenTest.ConsoleApp.csproj.approved.txt | 2 +- ...orTest.test.ConsoleApp.csproj.approved.txt | 2 +- ...CodeGenTest.ConsoleApp.csproj.approved.txt | 2 +- ...s.AzureImageCodeGeneratorTest.received.txt | 2 +- ...CodeGenTest.ConsoleApp.csproj.approved.txt | 2 +- ...soleAppProjectFileContentTest.approved.txt | 2 +- ...Contents_VerifyPredictProject.approved.txt | 2 +- .../RemoteExecutor.cs | 68 ------ .../UnitTests.cs | 159 +++++++------- test/Microsoft.ML.FSharp.Tests/SmokeTests.fs | 2 +- .../Microsoft.ML.TestFramework.csproj | 3 +- .../RemoteExecutor.cs | 204 ------------------ .../RemoteExecutorConsoleApp.cs | 124 ----------- .../RemoteExecutorConsoleApp.csproj | 8 - 21 files changed, 92 insertions(+), 514 deletions(-) delete mode 100644 test/Microsoft.ML.CpuMath.UnitTests/RemoteExecutor.cs delete mode 100644 test/Microsoft.ML.TestFramework/RemoteExecutor.cs delete mode 100644 test/RemoteExecutorConsoleApp/RemoteExecutorConsoleApp.cs delete mode 100644 test/RemoteExecutorConsoleApp/RemoteExecutorConsoleApp.csproj diff --git a/Directory.Build.props b/Directory.Build.props index 8e67d0905c..797eaa8a10 100644 --- a/Directory.Build.props +++ b/Directory.Build.props @@ -48,7 +48,7 @@ ML.NET ML Machine Learning - + true diff --git a/Microsoft.ML.sln b/Microsoft.ML.sln index 5763a903b4..61afe4c165 100644 --- a/Microsoft.ML.sln +++ b/Microsoft.ML.sln @@ -110,8 +110,6 @@ Project("{9A19103F-16F7-4668-BE54-9A1E7A4F7556}") = "Microsoft.ML.EntryPoints", EndProject Project("{9A19103F-16F7-4668-BE54-9A1E7A4F7556}") = "Microsoft.ML.DataView", "src\Microsoft.ML.DataView\Microsoft.ML.DataView.csproj", "{85D0CAFD-2FE8-496A-88C7-585D35B94243}" EndProject -Project("{9A19103F-16F7-4668-BE54-9A1E7A4F7556}") = "RemoteExecutorConsoleApp", "test\RemoteExecutorConsoleApp\RemoteExecutorConsoleApp.csproj", "{5E920CAC-5A28-42FB-936E-49C472130953}" -EndProject Project("{9A19103F-16F7-4668-BE54-9A1E7A4F7556}") = "Microsoft.ML.Experimental", "src\Microsoft.ML.Experimental\Microsoft.ML.Experimental.csproj", "{E02DA82D-3FEE-4C60-BD80-9EC3C3448DFC}" EndProject Project("{9A19103F-16F7-4668-BE54-9A1E7A4F7556}") = "Microsoft.Extensions.ML", "src\Microsoft.Extensions.ML\Microsoft.Extensions.ML.csproj", "{D6741C37-B5E6-4050-BCBA-9715809EA15B}" @@ -560,14 +558,6 @@ Global {85D0CAFD-2FE8-496A-88C7-585D35B94243}.Release|Any CPU.Build.0 = Release|Any CPU {85D0CAFD-2FE8-496A-88C7-585D35B94243}.Release|x64.ActiveCfg = Release|Any CPU {85D0CAFD-2FE8-496A-88C7-585D35B94243}.Release|x64.Build.0 = Release|Any CPU - {5E920CAC-5A28-42FB-936E-49C472130953}.Debug|Any CPU.ActiveCfg = Debug|Any CPU - {5E920CAC-5A28-42FB-936E-49C472130953}.Debug|Any CPU.Build.0 = Debug|Any CPU - {5E920CAC-5A28-42FB-936E-49C472130953}.Debug|x64.ActiveCfg = Debug|Any CPU - {5E920CAC-5A28-42FB-936E-49C472130953}.Debug|x64.Build.0 = Debug|Any CPU - {5E920CAC-5A28-42FB-936E-49C472130953}.Release|Any CPU.ActiveCfg = Release|Any CPU - {5E920CAC-5A28-42FB-936E-49C472130953}.Release|Any CPU.Build.0 = Release|Any CPU - {5E920CAC-5A28-42FB-936E-49C472130953}.Release|x64.ActiveCfg = Release|Any CPU - {5E920CAC-5A28-42FB-936E-49C472130953}.Release|x64.Build.0 = Release|Any CPU {E02DA82D-3FEE-4C60-BD80-9EC3C3448DFC}.Debug|Any CPU.ActiveCfg = Debug|Any CPU {E02DA82D-3FEE-4C60-BD80-9EC3C3448DFC}.Debug|Any CPU.Build.0 = Debug|Any CPU {E02DA82D-3FEE-4C60-BD80-9EC3C3448DFC}.Debug|x64.ActiveCfg = Debug|Any CPU @@ -880,7 +870,6 @@ Global {DB7CEB5E-8BE6-48A7-87BE-B91D9AE96F71} = {09EADF06-BE25-4228-AB53-95AE3E15B530} {7504D46F-E4B3-43CB-9B1C-82F3131F1C99} = {09EADF06-BE25-4228-AB53-95AE3E15B530} {85D0CAFD-2FE8-496A-88C7-585D35B94243} = {09EADF06-BE25-4228-AB53-95AE3E15B530} - {5E920CAC-5A28-42FB-936E-49C472130953} = {AED9C836-31E3-4F3F-8ABC-929555D3F3C4} {E02DA82D-3FEE-4C60-BD80-9EC3C3448DFC} = {09EADF06-BE25-4228-AB53-95AE3E15B530} {D6741C37-B5E6-4050-BCBA-9715809EA15B} = {09EADF06-BE25-4228-AB53-95AE3E15B530} {21CAD3A1-5E1F-42C1-BB73-46B6E67F4206} = {AED9C836-31E3-4F3F-8ABC-929555D3F3C4} diff --git a/docs/samples/Microsoft.ML.AutoML.Samples/Microsoft.ML.AutoML.Samples.csproj b/docs/samples/Microsoft.ML.AutoML.Samples/Microsoft.ML.AutoML.Samples.csproj index 628cbe5293..dca3c27c9b 100644 --- a/docs/samples/Microsoft.ML.AutoML.Samples/Microsoft.ML.AutoML.Samples.csproj +++ b/docs/samples/Microsoft.ML.AutoML.Samples/Microsoft.ML.AutoML.Samples.csproj @@ -2,7 +2,7 @@ Exe - net6.0 + net8.0 false $(NoWarn) diff --git a/docs/samples/Microsoft.ML.Samples/Microsoft.ML.Samples.csproj b/docs/samples/Microsoft.ML.Samples/Microsoft.ML.Samples.csproj index 8e168ad456..7e7442a029 100644 --- a/docs/samples/Microsoft.ML.Samples/Microsoft.ML.Samples.csproj +++ b/docs/samples/Microsoft.ML.Samples/Microsoft.ML.Samples.csproj @@ -1,7 +1,7 @@  - net6.0 + net8.0 Exe false diff --git a/eng/Version.Details.xml b/eng/Version.Details.xml index 48dbff16a5..b384be05ab 100644 --- a/eng/Version.Details.xml +++ b/eng/Version.Details.xml @@ -23,11 +23,10 @@ https://github.com/dotnet/arcade 812d978c303174dc1aa305d7359e79053d7d4971 - + https://github.com/dotnet/arcade 812d978c303174dc1aa305d7359e79053d7d4971 diff --git a/eng/Versions.props b/eng/Versions.props index 194e0211e1..44eefabb07 100644 --- a/eng/Versions.props +++ b/eng/Versions.props @@ -79,7 +79,7 @@ 8.0.1 5.10.2 1.1.2-beta1.23431.1 - 8.0.0-beta.24165.4 + 9.0.0-beta.24165.3 9.0.0-beta.24165.3 2.1.0 3.0.1 diff --git a/eng/helix.proj b/eng/helix.proj index effc0c5e3d..4dfb68d043 100644 --- a/eng/helix.proj +++ b/eng/helix.proj @@ -154,7 +154,7 @@ $(BUILD_SOURCESDIRECTORY)\artifacts\bin\%(ProjectsWithTargetFramework.Filename)\$(BuildConfig)\%(ProjectsWithTargetFramework.TargetFrameworks)\ $(BUILD_SOURCESDIRECTORY)\artifacts\bin\%(ProjectsWithTargetFramework.Filename)\$(BuildConfig)\%(ProjectsWithTargetFramework.TargetFrameworks)\win-x64 dotnet exec --roll-forward Major --runtimeconfig %(ProjectsWithTargetFramework.Filename).runtimeconfig.json --depsfile %(ProjectsWithTargetFramework.Filename).deps.json $(HelixCorrelationPayloadPath)/xunit-runner/tools/netcoreapp2.0/xunit.console.dll %(ProjectsWithTargetFramework.Filename).dll -notrait Category=SkipInCI -xml testResults.xml - $(HelixCorrelationPayloadPath)/xunit-runner/tools/net462/xunit.console.exe %(ProjectsWithTargetFramework.Filename).dll -notrait Category=SkipInCI -xml testResults.xml + $(HelixCorrelationPayloadPath)/xunit-runner/tools/net462/xunit.console.exe %(ProjectsWithTargetFramework.Filename).dll -notrait Category=SkipInCI -xml testResults.xml 01:00:00 00:30:00 $(WorkItemTimeout) diff --git a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt index d56e742f3b..b75a426ee8 100644 --- a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt +++ b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt @@ -2,7 +2,7 @@ Exe - net6.0 + net8.0 diff --git a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureCodeGeneratorTest.test.ConsoleApp.csproj.approved.txt b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureCodeGeneratorTest.test.ConsoleApp.csproj.approved.txt index 7e7c94d918..ef3297ad59 100644 --- a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureCodeGeneratorTest.test.ConsoleApp.csproj.approved.txt +++ b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureCodeGeneratorTest.test.ConsoleApp.csproj.approved.txt @@ -2,7 +2,7 @@ Exe - net6.0 + net8.0 diff --git a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureImageCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureImageCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt index 12a18142d2..9be23b26b9 100644 --- a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureImageCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt +++ b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureImageCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt @@ -2,7 +2,7 @@ Exe - net6.0 + net8.0 diff --git a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureImageCodeGeneratorTest.received.txt b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureImageCodeGeneratorTest.received.txt index 4637958288..27e748537e 100644 --- a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureImageCodeGeneratorTest.received.txt +++ b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureImageCodeGeneratorTest.received.txt @@ -2,7 +2,7 @@ Exe - net6.0 + net8.0 diff --git a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureObjectDetectionCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureObjectDetectionCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt index 12a18142d2..9be23b26b9 100644 --- a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureObjectDetectionCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt +++ b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.AzureObjectDetectionCodeGeneratorTest.CodeGenTest.ConsoleApp.csproj.approved.txt @@ -2,7 +2,7 @@ Exe - net6.0 + net8.0 diff --git a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.ConsoleAppProjectFileContentTest.approved.txt b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.ConsoleAppProjectFileContentTest.approved.txt index b6aaf2b243..ab60beee0e 100644 --- a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.ConsoleAppProjectFileContentTest.approved.txt +++ b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.ConsoleAppProjectFileContentTest.approved.txt @@ -2,7 +2,7 @@ Exe - net6.0 + net8.0 diff --git a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.Recommendation_GenerateConsoleAppProjectContents_VerifyPredictProject.approved.txt b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.Recommendation_GenerateConsoleAppProjectContents_VerifyPredictProject.approved.txt index 2ffb1661f0..d6141f8276 100644 --- a/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.Recommendation_GenerateConsoleAppProjectContents_VerifyPredictProject.approved.txt +++ b/test/Microsoft.ML.CodeGenerator.Tests/ApprovalTests/ConsoleCodeGeneratorTests.Recommendation_GenerateConsoleAppProjectContents_VerifyPredictProject.approved.txt @@ -2,7 +2,7 @@ Exe - net6.0 + net8.0 diff --git a/test/Microsoft.ML.CpuMath.UnitTests/RemoteExecutor.cs b/test/Microsoft.ML.CpuMath.UnitTests/RemoteExecutor.cs deleted file mode 100644 index f3e6e06e38..0000000000 --- a/test/Microsoft.ML.CpuMath.UnitTests/RemoteExecutor.cs +++ /dev/null @@ -1,68 +0,0 @@ -// Licensed to the .NET Foundation under one or more agreements. -// The .NET Foundation licenses this file to you under the MIT license. -// See the LICENSE file in the project root for more information. - -using System; - -#if NET6_0_OR_GREATER -using Executor = Microsoft.DotNet.RemoteExecutor.RemoteExecutor; -#else -using Executor = Microsoft.ML.TestFramework.RemoteExecutor; -#endif - -namespace Microsoft.ML.CpuMath.UnitTests -{ - - internal static class RemoteExecutor - { - public const int SuccessExitCode = 42; - - public static void RemoteInvoke( - Func method, - string arg1, string arg2, string arg3, string arg4, -#if NETFRAMEWORK - Microsoft.ML.TestFramework.RemoteInvokeOptions options = null) -#else - Microsoft.DotNet.RemoteExecutor.RemoteInvokeOptions options = null) -#endif - { -#if NETFRAMEWORK - Executor.RemoteInvoke(method, arg1, arg2, arg3, arg4, options); -#else - Executor.Invoke(method, arg1, arg2, arg3, arg4, options).Dispose(); -#endif - } - - public static void RemoteInvoke( - Func method, - string arg1, string arg2, string arg3, -#if NETFRAMEWORK - Microsoft.ML.TestFramework.RemoteInvokeOptions options = null) -#else - Microsoft.DotNet.RemoteExecutor.RemoteInvokeOptions options = null) -#endif - { -#if NETFRAMEWORK - Executor.RemoteInvoke(method, arg1, arg2, arg3, options); -#else - Executor.Invoke(method, arg1, arg2, arg3, options).Dispose(); -#endif - } - - public static void RemoteInvoke( - Func method, - string arg1, string arg2, -#if NETFRAMEWORK - Microsoft.ML.TestFramework.RemoteInvokeOptions options = null) -#else - Microsoft.DotNet.RemoteExecutor.RemoteInvokeOptions options = null) -#endif - { -#if NETFRAMEWORK - Executor.RemoteInvoke(method, arg1, arg2, options); -#else - Executor.Invoke(method, arg1, arg2, options).Dispose(); -#endif - } - } -} diff --git a/test/Microsoft.ML.CpuMath.UnitTests/UnitTests.cs b/test/Microsoft.ML.CpuMath.UnitTests/UnitTests.cs index edf99ef5c9..1146b5f01a 100644 --- a/test/Microsoft.ML.CpuMath.UnitTests/UnitTests.cs +++ b/test/Microsoft.ML.CpuMath.UnitTests/UnitTests.cs @@ -12,12 +12,7 @@ using Microsoft.ML.TestFramework; using Xunit; using Xunit.Abstractions; - -#if NETFRAMEWORK -using RemoteOptions = Microsoft.ML.TestFramework.RemoteInvokeOptions; -#else -using RemoteOptions = Microsoft.DotNet.RemoteExecutor.RemoteInvokeOptions; -#endif +using Microsoft.DotNet.RemoteExecutor; namespace Microsoft.ML.CpuMath.UnitTests { @@ -239,10 +234,10 @@ public static TheoryData environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2, arg3) => + RemoteExecutor.Invoke((arg0, arg1, arg2, arg3) => { CheckProperFlag(arg0); @@ -266,7 +261,7 @@ public void MatMulTest(string mode, string matTest, string srcTest, string dstTe dst.CopyTo(actual, 0, dst.Size); Assert.Equal(expected, actual, _matMulComparer); return RemoteExecutor.SuccessExitCode; - }, mode, matTest, srcTest, dstTest, options); + }, mode, matTest, srcTest, dstTest, options).Dispose(); } @@ -274,10 +269,10 @@ public void MatMulTest(string mode, string matTest, string srcTest, string dstTe [MemberData(nameof(MatMulData))] public void MatMulTranTest(string mode, string matTest, string srcTest, string dstTest, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2, arg3) => + RemoteExecutor.Invoke((arg0, arg1, arg2, arg3) => { CheckProperFlag(arg0); AlignedArray mat = _testMatrices[int.Parse(arg1)]; @@ -300,17 +295,17 @@ public void MatMulTranTest(string mode, string matTest, string srcTest, string d dst.CopyTo(actual, 0, dst.Size); Assert.Equal(expected, actual, _matMulComparer); return RemoteExecutor.SuccessExitCode; - }, mode, matTest, srcTest, dstTest, options); + }, mode, matTest, srcTest, dstTest, options).Dispose(); } [Theory] [MemberData(nameof(MatMulData))] public void MatTimesSrcSparseTest(string mode, string matTest, string srcTest, string dstTest, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2, arg3) => + RemoteExecutor.Invoke((arg0, arg1, arg2, arg3) => { CheckProperFlag(arg0); AlignedArray mat = _testMatrices[int.Parse(arg1)]; @@ -337,7 +332,7 @@ public void MatTimesSrcSparseTest(string mode, string matTest, string srcTest, s Assert.Equal(expected, actual, _matMulComparer); return RemoteExecutor.SuccessExitCode; - }, mode, matTest, srcTest, dstTest, options); + }, mode, matTest, srcTest, dstTest, options).Dispose(); } [Theory] @@ -345,10 +340,10 @@ public void MatTimesSrcSparseTest(string mode, string matTest, string srcTest, s [System.Diagnostics.CodeAnalysis.SuppressMessage("Usage", "xUnit1026:Theory methods should use all of their parameters", Justification = "")] public void AddScalarUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse(arg2, CultureInfo.InvariantCulture); @@ -364,17 +359,17 @@ public void AddScalarUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse(arg2, CultureInfo.InvariantCulture); @@ -390,17 +385,17 @@ public void ScaleTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse(arg2, CultureInfo.InvariantCulture); @@ -417,17 +412,17 @@ public void ScaleSrcUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse(arg2, CultureInfo.InvariantCulture); @@ -443,7 +438,7 @@ public void ScaleAddUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse(arg2, CultureInfo.InvariantCulture); @@ -471,17 +466,17 @@ public void AddScaleUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse(arg2, CultureInfo.InvariantCulture); @@ -500,17 +495,17 @@ public void AddScaleSUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse("1.7", CultureInfo.InvariantCulture); @@ -528,17 +523,17 @@ public void AddScaleCopyUTest(string mode, string test, string scale, Dictionary var actual = result; Assert.Equal(expected, actual, _comparer); return RemoteExecutor.SuccessExitCode; - }, mode, test, scale, options); + }, mode, test, scale, options).Dispose(); } [Theory] [MemberData(nameof(AddData))] public void AddUTest(string mode, string test, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1) => + RemoteExecutor.Invoke((arg0, arg1) => { CheckProperFlag(arg0); float[] src = (float[])_testArrays[int.Parse(arg1)].Clone(); @@ -560,7 +555,7 @@ public void AddUTest(string mode, string test, Dictionary enviro var actual = dst; Assert.Equal(expected, actual, _comparer); return RemoteExecutor.SuccessExitCode; - }, mode, test, options); + }, mode, test, options).Dispose(); } @@ -568,10 +563,10 @@ public void AddUTest(string mode, string test, Dictionary enviro [MemberData(nameof(AddData))] public void AddSUTest(string mode, string test, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1) => + RemoteExecutor.Invoke((arg0, arg1) => { CheckProperFlag(arg0); float[] src = (float[])_testArrays[int.Parse(arg1)].Clone(); @@ -590,17 +585,17 @@ public void AddSUTest(string mode, string test, Dictionary envir var actual = dst; Assert.Equal(expected, actual, _comparer); return RemoteExecutor.SuccessExitCode; - }, mode, test, options); + }, mode, test, options).Dispose(); } [Theory] [MemberData(nameof(AddData))] public void MulElementWiseUTest(string mode, string test, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1) => + RemoteExecutor.Invoke((arg0, arg1) => { CheckProperFlag(arg1); float[] src1 = (float[])_testArrays[int.Parse(arg1)].Clone(); @@ -624,17 +619,17 @@ public void MulElementWiseUTest(string mode, string test, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1) => + RemoteExecutor.Invoke((arg0, arg1) => { CheckProperFlag(arg0); float[] src = (float[])_testArrays[int.Parse(arg1)].Clone(); @@ -647,17 +642,17 @@ public void SumTest(string mode, string test, Dictionary environ var actual = CpuMathUtils.Sum(src); Assert.Equal((double)expected, (double)actual, 0.01); return RemoteExecutor.SuccessExitCode; - }, mode, test, options); + }, mode, test, options).Dispose(); } [Theory] [MemberData(nameof(AddData))] public void SumSqUTest(string mode, string test, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1) => + RemoteExecutor.Invoke((arg0, arg1) => { CheckProperFlag(arg0); float[] src = (float[])_testArrays[int.Parse(arg1)].Clone(); @@ -670,17 +665,17 @@ public void SumSqUTest(string mode, string test, Dictionary envi var actual = CpuMathUtils.SumSq(src); Assert.Equal((double)expected, (double)actual, 0.01); return RemoteExecutor.SuccessExitCode; - }, mode, test, options); + }, mode, test, options).Dispose(); } [Theory] [MemberData(nameof(AddScaleData))] public void SumSqDiffUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse(arg2, CultureInfo.InvariantCulture); @@ -695,17 +690,17 @@ public void SumSqDiffUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1) => + RemoteExecutor.Invoke((arg0, arg1) => { CheckProperFlag(arg0); float[] src = (float[])_testArrays[int.Parse(arg1)].Clone(); @@ -718,17 +713,17 @@ public void SumAbsUTest(string mode, string test, Dictionary env var actual = CpuMathUtils.SumAbs(src); Assert.Equal((double)expected, (double)actual, 0.01); return RemoteExecutor.SuccessExitCode; - }, mode, test, options); + }, mode, test, options).Dispose(); } [Theory] [MemberData(nameof(AddScaleData))] public void SumAbsDiffUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse(arg2, CultureInfo.InvariantCulture); @@ -743,17 +738,17 @@ public void SumAbsDiffUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1) => + RemoteExecutor.Invoke((arg0, arg1) => { CheckProperFlag(arg0); float[] src = (float[])_testArrays[int.Parse(arg1)].Clone(); @@ -771,17 +766,17 @@ public void MaxAbsUTest(string mode, string test, Dictionary env Assert.Equal((double)expected, (double)actual, 0.01); return RemoteExecutor.SuccessExitCode; - }, mode, test, options); + }, mode, test, options).Dispose(); } [Theory] [MemberData(nameof(AddScaleData))] public void MaxAbsDiffUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse(arg2, CultureInfo.InvariantCulture); @@ -799,17 +794,17 @@ public void MaxAbsDiffUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1) => + RemoteExecutor.Invoke((arg0, arg1) => { CheckProperFlag(arg0); float[] src = (float[])_testArrays[int.Parse(arg1)].Clone(); @@ -829,17 +824,17 @@ public void DotUTest(string mode, string test, Dictionary enviro var actual = CpuMathUtils.DotProductDense(src, dst, dst.Length); Assert.Equal((double)expected, (double)actual, 0.1); return RemoteExecutor.SuccessExitCode; - }, mode, test, options); + }, mode, test, options).Dispose(); } [Theory] [MemberData(nameof(AddData))] public void DotSUTest(string mode, string test, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1) => + RemoteExecutor.Invoke((arg0, arg1) => { CheckProperFlag(arg0); float[] src = (float[])_testArrays[int.Parse(arg1)].Clone(); @@ -863,17 +858,17 @@ public void DotSUTest(string mode, string test, Dictionary envir var actual = CpuMathUtils.DotProductSparse(src, dst, idx, limit); Assert.Equal((double)expected, (double)actual, 0.01); return RemoteExecutor.SuccessExitCode; - }, mode, test, options); + }, mode, test, options).Dispose(); } [Theory] [MemberData(nameof(AddData))] public void Dist2Test(string mode, string test, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1) => + RemoteExecutor.Invoke((arg0, arg1) => { CheckProperFlag(arg0); float[] src = (float[])_testArrays[int.Parse(arg1)].Clone(); @@ -895,7 +890,7 @@ public void Dist2Test(string mode, string test, Dictionary envir var actual = CpuMathUtils.L2DistSquared(src, dst, dst.Length); Assert.Equal((double)expected, (double)actual, 0); return RemoteExecutor.SuccessExitCode; - }, mode, test, options); + }, mode, test, options).Dispose(); } [Theory] @@ -930,10 +925,10 @@ public void ZeroMatrixItemsCoreTest(int test, int[] idx, float[] expected) [MemberData(nameof(AddScaleData))] public void SdcaL1UpdateUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse(arg2, CultureInfo.InvariantCulture); @@ -952,7 +947,7 @@ public void SdcaL1UpdateUTest(string mode, string test, string scale, Dictionary var actual = w; Assert.Equal(expected, actual, _comparer); return RemoteExecutor.SuccessExitCode; - }, mode, test, scale, options); + }, mode, test, scale, options).Dispose(); } @@ -960,10 +955,10 @@ public void SdcaL1UpdateUTest(string mode, string test, string scale, Dictionary [MemberData(nameof(AddScaleData))] public void SdcaL1UpdateSUTest(string mode, string test, string scale, Dictionary environmentVariables) { - var options = new RemoteOptions(); + var options = new RemoteInvokeOptions(); UpdateEnvVars(options, environmentVariables); - RemoteExecutor.RemoteInvoke((arg0, arg1, arg2) => + RemoteExecutor.Invoke((arg0, arg1, arg2) => { CheckProperFlag(arg0); float defaultScale = float.Parse(arg2, CultureInfo.InvariantCulture); @@ -985,10 +980,10 @@ public void SdcaL1UpdateSUTest(string mode, string test, string scale, Dictionar var actual = w; Assert.Equal(expected, actual, _comparer); return RemoteExecutor.SuccessExitCode; - }, mode, test, scale, options); + }, mode, test, scale, options).Dispose(); } - private void UpdateEnvVars(RemoteOptions options, Dictionary environmentVariables) + private void UpdateEnvVars(RemoteInvokeOptions options, Dictionary environmentVariables) { if (environmentVariables == null) return; diff --git a/test/Microsoft.ML.FSharp.Tests/SmokeTests.fs b/test/Microsoft.ML.FSharp.Tests/SmokeTests.fs index 0f129eacfc..c2b5be314e 100644 --- a/test/Microsoft.ML.FSharp.Tests/SmokeTests.fs +++ b/test/Microsoft.ML.FSharp.Tests/SmokeTests.fs @@ -14,7 +14,7 @@ // You should build Microsoft.ML.FSharp.Tests in Debug mode for framework net462 // before running this as a script with F# Interactive by editing the project // file to have: -// net6.0; net462 +// net8.0; net462 #if INTERACTIVE #r "netstandard" diff --git a/test/Microsoft.ML.TestFramework/Microsoft.ML.TestFramework.csproj b/test/Microsoft.ML.TestFramework/Microsoft.ML.TestFramework.csproj index 483cb1ab7a..53783ee35b 100644 --- a/test/Microsoft.ML.TestFramework/Microsoft.ML.TestFramework.csproj +++ b/test/Microsoft.ML.TestFramework/Microsoft.ML.TestFramework.csproj @@ -19,7 +19,6 @@ - @@ -33,7 +32,7 @@ - + diff --git a/test/Microsoft.ML.TestFramework/RemoteExecutor.cs b/test/Microsoft.ML.TestFramework/RemoteExecutor.cs deleted file mode 100644 index 097de7ad8e..0000000000 --- a/test/Microsoft.ML.TestFramework/RemoteExecutor.cs +++ /dev/null @@ -1,204 +0,0 @@ -// Licensed to the .NET Foundation under one or more agreements. -// The .NET Foundation licenses this file to you under the MIT license. -// See the LICENSE file in the project root for more information. - -using System; -using System.Collections.Generic; -using System.Diagnostics; -using System.IO; -using System.Reflection; -using System.Runtime.InteropServices; -using System.Threading.Tasks; -using Xunit; -using Xunit.Sdk; - -namespace Microsoft.ML.TestFramework -{ - /// - /// Base class used for all tests that need to spawn a remote process. - /// Most of the code has been taken from RemoteExecutorTestBase class in the corefx repo. - /// - public static class RemoteExecutor - { - /// The name of the test console app. - public static readonly string TestConsoleApp = Path.GetFullPath(@"RemoteExecutorConsoleApp.dll"); -#if NETFRAMEWORK - public static readonly string HostRunner = Path.GetFullPath(@"RemoteExecutorConsoleApp.exe"); - private static readonly string _extraParameter = ""; -#else - public static readonly string HostRunner = Process.GetCurrentProcess().MainModule.FileName; - private static readonly string _extraParameter = TestConsoleApp; -#endif - /// A timeout (milliseconds) after which a wait on a remote operation should be considered a failure. - public const int FailWaitTimeoutMilliseconds = 60 * 1000; - - /// The exit code returned when the test process exits successfully. - public const int SuccessExitCode = 42; - - /// Invokes the method from this assembly in another process using the specified arguments. - /// The method to invoke. - /// The first argument to pass to the method. - /// The second argument to pass to the method. - /// Options to use for the invocation. - public static void RemoteInvoke( - Func method, - string arg1, string arg2, - RemoteInvokeOptions options = null) - { - RemoteInvoke(GetMethodInfo(method), new[] { arg1, arg2 }, options); - } - - /// Invokes the method from this assembly in another process using the specified arguments. - /// The method to invoke. - /// The first argument to pass to the method. - /// The second argument to pass to the method. - /// The third argument to pass to the method. - /// Options to use for the invocation. - public static void RemoteInvoke( - Func method, - string arg1, string arg2, string arg3, - RemoteInvokeOptions options = null) - { - RemoteInvoke(GetMethodInfo(method), new[] { arg1, arg2, arg3 }, options); - } - - /// Invokes the method from this assembly in another process using the specified arguments. - /// The method to invoke. - /// The first argument to pass to the method. - /// The second argument to pass to the method. - /// The third argument to pass to the method. - /// The fourth argument to pass to the method. - /// Options to use for the invocation. - public static void RemoteInvoke( - Func method, - string arg1, string arg2, string arg3, string arg4, - RemoteInvokeOptions options = null) - { - RemoteInvoke(GetMethodInfo(method), new[] { arg1, arg2, arg3, arg4 }, options); - } - - /// Invokes the method from this assembly in another process using the specified arguments. - /// The method to invoke. - /// The arguments to pass to the method. - /// Options to use for the invocation. - /// true if this function should paste the arguments (e.g. surrounding with quotes); false if that responsibility is left up to the caller. - private static void RemoteInvoke(MethodInfo method, string[] args, RemoteInvokeOptions options, bool pasteArguments = true) - { - options = options ?? new RemoteInvokeOptions(); - - // Verify the specified method returns an int (the exit code) or nothing, - // and that if it accepts any arguments, they're all strings. - Assert.True(method.ReturnType == typeof(void) || method.ReturnType == typeof(int) || method.ReturnType == typeof(Task)); - Assert.All(method.GetParameters(), pi => Assert.Equal(typeof(string), pi.ParameterType)); - - // And make sure it's in this assembly. This isn't critical, but it helps with deployment to know - // that the method to invoke is available because we're already running in this assembly. - Type t = method.DeclaringType; - Assembly a = t.GetTypeInfo().Assembly; - - // Start the other process and return a wrapper for it to handle its lifetime and exit checking. - ProcessStartInfo psi = options.StartInfo; - psi.UseShellExecute = false; - - // If we need the host (if it exists), use it, otherwise target the console app directly. - string metadataArgs = PasteArguments.Paste(new string[] { a.FullName, t.FullName, method.Name, options.ExceptionFile }, pasteFirstArgumentUsingArgV0Rules: false); - string passedArgs = pasteArguments ? PasteArguments.Paste(args, pasteFirstArgumentUsingArgV0Rules: false) : string.Join(" ", args); - string testConsoleAppArgs = _extraParameter + " " + metadataArgs + " " + passedArgs; - - psi.FileName = HostRunner; - psi.Arguments = testConsoleAppArgs; - - // Return the handle to the process, which may or not be started - CheckProcess(Process.Start(psi), options); - } - - private static void CheckProcess(Process process, RemoteInvokeOptions options) - { - if (process != null) - { - // A bit unorthodox to do throwing operations in a Dispose, but by doing it here we avoid - // needing to do this in every derived test and keep each test much simpler. - try - { - Assert.True(process.WaitForExit(options.TimeOut), - $"Timed out after {options.TimeOut}ms waiting for remote process {process.Id}"); - - if (File.Exists(options.ExceptionFile)) - { - throw new RemoteExecutionException(File.ReadAllText(options.ExceptionFile)); - } - - if (options.CheckExitCode) - { - int expected = RuntimeInformation.IsOSPlatform(OSPlatform.Windows) ? options.ExpectedExitCode : unchecked((sbyte)options.ExpectedExitCode); - int actual = RuntimeInformation.IsOSPlatform(OSPlatform.Windows) ? process.ExitCode : unchecked((sbyte)process.ExitCode); - - Assert.True(expected == actual, $"Exit code was {process.ExitCode} but it should have been {options.ExpectedExitCode}"); - } - } - finally - { - if (File.Exists(options.ExceptionFile)) - { - File.Delete(options.ExceptionFile); - } - - // Cleanup - try { process.Kill(); } - catch { } // ignore all cleanup errors - - process.Dispose(); - process = null; - } - } - } - - private sealed class RemoteExecutionException : XunitException - { - internal RemoteExecutionException(string stackTrace) : base($"Remote process failed with an unhandled exception. {stackTrace}") { } - } - - private static MethodInfo GetMethodInfo(Delegate d) - { - // RemoteInvoke doesn't support marshaling state on classes associated with - // the delegate supplied (often a display class of a lambda). If such fields - // are used, odd errors result, e.g. NullReferenceExceptions during the remote - // execution. Try to ward off the common cases by proactively failing early - // if it looks like such fields are needed. - if (d.Target != null) - { - // The only fields on the type should be compiler-defined (any fields of the compiler's own - // making generally include '<' and '>', as those are invalid in C# source). Note that this logic - // may need to be revised in the future as the compiler changes, as this relies on the specifics of - // actually how the compiler handles lifted fields for lambdas. - Type targetType = d.Target.GetType(); - Assert.All( - targetType.GetFields(BindingFlags.Public | BindingFlags.NonPublic | BindingFlags.Instance | BindingFlags.DeclaredOnly), - fi => Assert.True(fi.Name.IndexOf('<') != -1, $"Field marshaling is not supported by {nameof(RemoteInvoke)}: {fi.Name}")); - } - - return d.GetMethodInfo(); - } - } - - /// Options used with RemoteInvoke. - public sealed class RemoteInvokeOptions - { - public RemoteInvokeOptions(Dictionary environmentVariables = null) - { - if (environmentVariables != null) - { - foreach (var item in environmentVariables) - { - StartInfo.EnvironmentVariables.Add(item.Key, item.Value); - } - } - } - - public ProcessStartInfo StartInfo { get; set; } = new ProcessStartInfo(); - public bool CheckExitCode { get; set; } = true; - public int TimeOut { get; set; } = RemoteExecutor.FailWaitTimeoutMilliseconds; - public int ExpectedExitCode { get; set; } = RemoteExecutor.SuccessExitCode; - public string ExceptionFile { get; } = Path.Combine(Path.GetTempPath(), Path.GetRandomFileName()); - } -} diff --git a/test/RemoteExecutorConsoleApp/RemoteExecutorConsoleApp.cs b/test/RemoteExecutorConsoleApp/RemoteExecutorConsoleApp.cs deleted file mode 100644 index 113ee6a4c1..0000000000 --- a/test/RemoteExecutorConsoleApp/RemoteExecutorConsoleApp.cs +++ /dev/null @@ -1,124 +0,0 @@ -// Licensed to the .NET Foundation under one or more agreements. -// The .NET Foundation licenses this file to you under the MIT license. -// See the LICENSE file in the project root for more information. - -using System; -using System.Diagnostics; -using System.IO; -using System.Reflection; -using System.Runtime.ExceptionServices; -using System.Runtime.InteropServices; -using System.Text; -using System.Threading.Tasks; - -namespace RemoteExecutorConsoleApp -{ - class Program - { - static int Main(string[] args) - { - if (args.Length < 4) - { - Console.Error.WriteLine("Usage: {0} assemblyName typeName methodName exceptionFile [additionalArgs]", typeof(Program).GetTypeInfo().Assembly.GetName().Name); - Environment.Exit(-1); - return -1; - } - - string assemblyName = args[0]; - string typeName = args[1]; - string methodName = args[2]; - string exceptionFile = args[3]; - string[] additionalArgs = args.Length > 4 ? - Subarray(args, 4, args.Length - 4) : - Array.Empty(); - - // Load the specified assembly, type, and method, then invoke the method. - // The program's exit code is the return value of the invoked method. - Assembly a = null; - Type t = null; - MethodInfo mi = null; - object instance = null; - int exitCode = 0; - - try - { - // Create the test class if necessary - try - { - a = Assembly.Load(assemblyName); - } - catch (FileNotFoundException) - { - a = Assembly.LoadFrom(assemblyName.Split(',')[0] + ".dll"); - } - - t = a.GetType(typeName); - mi = t.GetTypeInfo().GetDeclaredMethod(methodName); - if (!mi.IsStatic) - { - instance = Activator.CreateInstance(t); - } - - // Invoke the test - object result = mi.Invoke(instance, additionalArgs); - - if (result is Task task) - { - exitCode = task.GetAwaiter().GetResult(); - } - else if (result is int exit) - { - exitCode = exit; - } - } - catch (Exception exc) - { - if (exc is TargetInvocationException && exc.InnerException != null) - exc = exc.InnerException; - - var output = new StringBuilder(); - output.AppendLine(); - output.AppendLine("Child exception:"); - output.AppendLine(" " + exc); - output.AppendLine(); - output.AppendLine("Child process:"); - output.AppendLine(string.Format(" {0} {1} {2}", a, t, mi)); - output.AppendLine(); - - if (additionalArgs.Length > 0) - { - output.AppendLine("Child arguments:"); - output.AppendLine(" " + string.Join(", ", additionalArgs)); - } - - File.WriteAllText(exceptionFile, output.ToString()); - - ExceptionDispatchInfo.Capture(exc).Throw(); - } - finally - { - (instance as IDisposable)?.Dispose(); - } - - // Use Exit rather than simply returning the exit code so that we forcibly shut down - // the process even if there are foreground threads created by the operation that would - // end up keeping the process alive potentially indefinitely. - try - { - Environment.Exit(exitCode); - } - catch (PlatformNotSupportedException) - { - } - - return exitCode; - } - - private static T[] Subarray(T[] arr, int offset, int count) - { - var newArr = new T[count]; - Array.Copy(arr, offset, newArr, 0, count); - return newArr; - } - } -} diff --git a/test/RemoteExecutorConsoleApp/RemoteExecutorConsoleApp.csproj b/test/RemoteExecutorConsoleApp/RemoteExecutorConsoleApp.csproj deleted file mode 100644 index 580d921f67..0000000000 --- a/test/RemoteExecutorConsoleApp/RemoteExecutorConsoleApp.csproj +++ /dev/null @@ -1,8 +0,0 @@ - - - - Exe - false - - - From e18f8b1c3bf8dc4d3a100d72ebb7e41350bb4aea Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Thu, 21 Mar 2024 12:33:31 -0700 Subject: [PATCH 03/26] Remove net6.0 download --- eng/Versions.props | 1 - global.json | 2 -- 2 files changed, 3 deletions(-) diff --git a/eng/Versions.props b/eng/Versions.props index 44eefabb07..27f26cf54a 100644 --- a/eng/Versions.props +++ b/eng/Versions.props @@ -75,7 +75,6 @@ 5.4.7 0.13.1 - 6.0.26 8.0.1 5.10.2 1.1.2-beta1.23431.1 diff --git a/global.json b/global.json index 82ecb63c8f..cc537d7ae9 100644 --- a/global.json +++ b/global.json @@ -3,11 +3,9 @@ "dotnet": "9.0.100-preview.1.24101.2", "runtimes": { "dotnet": [ - "$(DotNetRuntime60Version)", "$(DotNetRuntime80Version)" ], "dotnet/x86": [ - "$(DotNetRuntime60Version)", "$(DotNetRuntime80Version)" ] } From e84dc7f16e3097cb90da365e002e69b8af070443 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Thu, 21 Mar 2024 12:35:37 -0700 Subject: [PATCH 04/26] Address change in behavior in enum.ToString InternalDataKind.TX and InternalDataKind.Text have the same underlying value. In net8.0 we started seeing ToString return the latter. --- .../CommandLine/CmdParser.cs | 5 ++++ .../Common/EntryPoints/core_manifest.json | 26 +++++++++---------- 2 files changed, 18 insertions(+), 13 deletions(-) diff --git a/src/Microsoft.ML.Core/CommandLine/CmdParser.cs b/src/Microsoft.ML.Core/CommandLine/CmdParser.cs index fc97810663..d698877600 100644 --- a/src/Microsoft.ML.Core/CommandLine/CmdParser.cs +++ b/src/Microsoft.ML.Core/CommandLine/CmdParser.cs @@ -10,6 +10,7 @@ using System.Linq; using System.Reflection; using System.Text; +using Microsoft.ML.Data; using Microsoft.ML.Internal.Utilities; using Microsoft.ML.Runtime; @@ -2015,6 +2016,10 @@ private void AppendSettingCore(IHostEnvironment env, StringBuilder sb, object va var buffer = new StringBuilder(); sb.Append(GetString(env, value, buffer)); } + else if (value is InternalDataKind) + { + sb.Append(((InternalDataKind)value).GetString()); + } else sb.Append(value.ToString()); } diff --git a/test/BaselineOutput/Common/EntryPoints/core_manifest.json b/test/BaselineOutput/Common/EntryPoints/core_manifest.json index 3dbaad5989..fe0fc65c73 100644 --- a/test/BaselineOutput/Common/EntryPoints/core_manifest.json +++ b/test/BaselineOutput/Common/EntryPoints/core_manifest.json @@ -173,13 +173,13 @@ "R4", "Num", "R8", - "TX", "Text", + "TX", "TXT", "BL", "Bool", - "TimeSpan", "TS", + "TimeSpan", "DT", "DateTime", "DZ", @@ -18160,13 +18160,13 @@ "R4", "Num", "R8", - "TX", "Text", + "TX", "TXT", "BL", "Bool", - "TimeSpan", "TS", + "TimeSpan", "DT", "DateTime", "DZ", @@ -18255,13 +18255,13 @@ "R4", "Num", "R8", - "TX", "Text", + "TX", "TXT", "BL", "Bool", - "TimeSpan", "TS", + "TimeSpan", "DT", "DateTime", "DZ", @@ -21779,7 +21779,7 @@ "Required": false, "SortOrder": 2.0, "IsNullable": false, - "Default": "Def" + "Default": "DefaultValue" }, { "Name": "ImputeBySlot", @@ -22150,7 +22150,7 @@ "Required": false, "SortOrder": 150.0, "IsNullable": false, - "Default": "Def" + "Default": "DefaultValue" }, { "Name": "ImputeBySlot", @@ -30296,13 +30296,13 @@ "R4", "Num", "R8", - "TX", "Text", + "TX", "TXT", "BL", "Bool", - "TimeSpan", "TS", + "TimeSpan", "DT", "DateTime", "DZ", @@ -30354,13 +30354,13 @@ "R4", "Num", "R8", - "TX", "Text", + "TX", "TXT", "BL", "Bool", - "TimeSpan", "TS", + "TimeSpan", "DT", "DateTime", "DZ", @@ -30373,7 +30373,7 @@ "Required": false, "SortOrder": 150.0, "IsNullable": false, - "Default": "TX" + "Default": "Text" } ] } From aea67f3c83c23d118125073d381ab74ec0d05bc9 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Thu, 21 Mar 2024 13:06:28 -0700 Subject: [PATCH 05/26] Update pipeline names --- .vsts-dotnet-ci.yml | 10 +++++----- README.ko-KR.md | 12 ++++++------ README.md | 12 ++++++------ build/.night-build.yml | 4 ++-- build/.outer-loop-build.yml | 12 ++++++------ 5 files changed, 25 insertions(+), 25 deletions(-) diff --git a/.vsts-dotnet-ci.yml b/.vsts-dotnet-ci.yml index a9a62df728..485b8fb2f1 100644 --- a/.vsts-dotnet-ci.yml +++ b/.vsts-dotnet-ci.yml @@ -43,7 +43,7 @@ jobs: - template: /build/ci/job-template.yml parameters: architecture: arm - name: Ubuntu_x64_cross_arm_Net60 + name: Ubuntu_x64_cross_arm_Net80 buildScript: ./build.sh container: UbuntuCrossArmContainer customMatrixes: @@ -66,7 +66,7 @@ jobs: - template: /build/ci/job-template.yml parameters: architecture: arm64 - name: Ubuntu_x64_cross_arm64_Net60 + name: Ubuntu_x64_cross_arm64_Net80 buildScript: ./build.sh container: UbuntuCrossArm64Container customMatrixes: @@ -88,7 +88,7 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: Centos_x64_Net60 + name: Centos_x64_Net80 buildScript: ./build.sh container: CentosContainer innerLoop: true @@ -99,7 +99,7 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: Ubuntu_x64_Net60 + name: Ubuntu_x64_Net80 buildScript: ./build.sh container: UbuntuContainer innerLoop: true @@ -110,7 +110,7 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: MacOS_x64_Net60 + name: MacOS_x64_Net80 buildScript: ./build.sh innerLoop: true pool: diff --git a/README.ko-KR.md b/README.ko-KR.md index f1d0507893..51c87713d8 100644 --- a/README.ko-KR.md +++ b/README.ko-KR.md @@ -70,13 +70,13 @@ Install-Package Microsoft.ML | | Debug | Release | |:---|----------------:|------------------:| -|**CentOS**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Centos_x64_Net60&configuration=Centos_x64_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Centos_x64_Net60&configuration=Centos_x64_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| -|**Ubuntu**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Ubuntu_x64_Net60&configuration=Ubuntu_x64_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Ubuntu_x64_Net60&configuration=Ubuntu_x64_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| -|**macOS**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=MacOS_x64_Net60&configuration=MacOS_x64_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=MacOS_x64_Net60&configuration=MacOS_x64_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| -|**Windows x64**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net60&configuration=Windows_x64_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net60&configuration=Windows_x64_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**CentOS**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Centos_x64_Net80&configuration=Centos_x64_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Centos_x64_Net80&configuration=Centos_x64_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**Ubuntu**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Ubuntu_x64_Net80&configuration=Ubuntu_x64_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Ubuntu_x64_Net80&configuration=Ubuntu_x64_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**macOS**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=MacOS_x64_Net80&configuration=MacOS_x64_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=MacOS_x64_Net80&configuration=MacOS_x64_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**Windows x64**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net80&configuration=Windows_x64_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net80&configuration=Windows_x64_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| |**Windows FullFramework**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_NetFx461&configuration=Windows_x64_NetFx461%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_NetFx461&configuration=Windows_x64_NetFx461%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| -|**Windows x86**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x86_Net60&configuration=Windows_x86_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x86_Net60&configuration=Windows_x86_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| -|**Windows NetCore3.1**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net60&configuration=Windows_x64_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net60&configuration=Windows_x64_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**Windows x86**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x86_Net80&configuration=Windows_x86_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x86_Net80&configuration=Windows_x86_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**Windows NetCore3.1**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net80&configuration=Windows_x64_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net80&configuration=Windows_x64_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| ## 릴리즈 프로세스 및 버전 관리 diff --git a/README.md b/README.md index b2dd1b7a1c..fd8bc9bf8a 100644 --- a/README.md +++ b/README.md @@ -66,13 +66,13 @@ To build ML.NET from source please visit our [developer guide](docs/project-docs | | Debug | Release | |:---|----------------:|------------------:| -|**CentOS**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Centos_x64_Net60&configuration=Centos_x64_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Centos_x64_Net60&configuration=Centos_x64_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| -|**Ubuntu**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Ubuntu_x64_Net60&configuration=Ubuntu_x64_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Ubuntu_x64_Net60&configuration=Ubuntu_x64_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| -|**macOS**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=MacOS_x64_Net60&configuration=MacOS_x64_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=MacOS_x64_Net60&configuration=MacOS_x64_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| -|**Windows x64**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net60&configuration=Windows_x64_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net60&configuration=Windows_x64_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**CentOS**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Centos_x64_Net80&configuration=Centos_x64_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Centos_x64_Net80&configuration=Centos_x64_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**Ubuntu**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Ubuntu_x64_Net80&configuration=Ubuntu_x64_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Ubuntu_x64_Net80&configuration=Ubuntu_x64_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**macOS**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=MacOS_x64_Net80&configuration=MacOS_x64_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=MacOS_x64_Net80&configuration=MacOS_x64_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**Windows x64**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net80&configuration=Windows_x64_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net80&configuration=Windows_x64_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| |**Windows FullFramework**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_NetFx461&configuration=Windows_x64_NetFx461%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_NetFx461&configuration=Windows_x64_NetFx461%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| -|**Windows x86**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x86_Net60&configuration=Windows_x86_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x86_Net60&configuration=Windows_x86_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| -|**Windows NetCore3.1**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net60&configuration=Windows_x64_Net60%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net60&configuration=Windows_x64_Net60%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**Windows x86**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x86_Net80&configuration=Windows_x86_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x86_Net80&configuration=Windows_x86_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| +|**Windows NetCore3.1**|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net80&configuration=Windows_x64_Net80%20Debug_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)|[![Build Status](https://dev.azure.com/dnceng/public/_apis/build/status/dotnet/machinelearning/MachineLearning-CI?branchName=main&jobName=Windows_x64_Net80&configuration=Windows_x64_Net80%20Release_Build)](https://dev.azure.com/dnceng/public/_build/latest?definitionId=104&branchName=main)| ## Release process and versioning diff --git a/build/.night-build.yml b/build/.night-build.yml index 6f2f8cf512..6f4479d0ba 100644 --- a/build/.night-build.yml +++ b/build/.night-build.yml @@ -30,7 +30,7 @@ resources: jobs: - template: /build/ci/job-template.yml parameters: - name: Centos_x64_Net60 + name: Centos_x64_Net80 buildScript: ./build.sh container: CentosContainer customMatrixes: @@ -67,7 +67,7 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: Windows_x64_Net60 + name: Windows_x64_Net80 buildScript: build.cmd customMatrixes: Debug_Build: diff --git a/build/.outer-loop-build.yml b/build/.outer-loop-build.yml index 2c58e6aacd..e96dab3bd4 100644 --- a/build/.outer-loop-build.yml +++ b/build/.outer-loop-build.yml @@ -31,7 +31,7 @@ resources: jobs: - template: /build/ci/job-template.yml parameters: - name: Centos_x64_Net60 + name: Centos_x64_Net80 buildScript: ./build.sh container: CentosContainer customMatrixes: @@ -48,7 +48,7 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: Ubuntu_x64_Net60 + name: Ubuntu_x64_Net80 buildScript: ./build.sh container: UbuntuContainer pool: @@ -56,14 +56,14 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: MacOS_x64_Net60 + name: MacOS_x64_Net80 buildScript: ./build.sh pool: vmImage: macOS-10.15 - template: /build/ci/job-template.yml parameters: - name: Windows_x64_Net60 + name: Windows_x64_Net80 buildScript: build.cmd customMatrixes: Debug_Build: @@ -79,7 +79,7 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: Windows_x64_Net60 + name: Windows_x64_Net80 buildScript: build.cmd pool: vmImage: windows-2019 @@ -104,7 +104,7 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: Windows_x86_Net60 + name: Windows_x86_Net80 architecture: x86 buildScript: build.cmd pool: From 4fe2869d8dbce1a389efebceb510c8d4404c9f0b Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Thu, 21 Mar 2024 13:08:42 -0700 Subject: [PATCH 06/26] Honor TestTargetFramework in tests. --- .../Microsoft.Data.Analysis.Interactive.Tests.csproj | 2 +- .../Microsoft.Data.Analysis.PerformanceTests.csproj | 2 +- .../Microsoft.ML.CpuMath.PerformanceTests.csproj | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj b/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj index 44f6d29bc7..f9b4850bd1 100644 --- a/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj +++ b/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj @@ -1,7 +1,7 @@  - $(NetMinimum);$(NetCurrent) + $(NetMinimum);$(NetCurrent) $(NoWarn);MSML_ExtendBaseTestClass diff --git a/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj b/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj index 39e6f38efa..dadf551d52 100644 --- a/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj +++ b/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj @@ -3,7 +3,7 @@ Exe disable - $(NetMinimum);$(NetCurrent) + $(NetMinimum);$(NetCurrent) false diff --git a/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj b/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj index 802c72c729..069a75d4ad 100644 --- a/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj +++ b/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj @@ -3,7 +3,7 @@ Exe false - $(NetMinimum);$(NetCurrent) + $(NetMinimum);$(NetCurrent) true false true From 194885b68f2bd362a134ffcb4127a1b5f456d088 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Thu, 21 Mar 2024 16:50:17 -0700 Subject: [PATCH 07/26] Update helix.proj --- eng/helix.proj | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/eng/helix.proj b/eng/helix.proj index 4dfb68d043..7a3dc0f17e 100644 --- a/eng/helix.proj +++ b/eng/helix.proj @@ -39,7 +39,7 @@ - + runtime win-x86 linux-arm @@ -161,4 +161,4 @@ - \ No newline at end of file + From 704fd6108dfbd2b27d1ae84a85a74f1a5b548ab0 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Fri, 22 Mar 2024 13:21:39 -0700 Subject: [PATCH 08/26] Remove build pipelines that don't run --- build/.night-build.yml | 119 --------------------------------- build/.outer-loop-build.yml | 111 ------------------------------- build/ci/job-template.yml | 127 +++++++++++++----------------------- 3 files changed, 47 insertions(+), 310 deletions(-) delete mode 100644 build/.night-build.yml delete mode 100644 build/.outer-loop-build.yml diff --git a/build/.night-build.yml b/build/.night-build.yml deleted file mode 100644 index 6f4479d0ba..0000000000 --- a/build/.night-build.yml +++ /dev/null @@ -1,119 +0,0 @@ -################################################################################ -# ML.NET's nightly build -################################################################################ - -# no PR builds -pr: none - -# no CI builds -trigger: none - -# scheduled trigger, runs at UTC 8:00 every day which is midnight of GMT-8 -schedules: -- cron: "0 8 * * *" - displayName: Nightly Build at midnight - branches: - include: - - main - - feature/* - - release/* - always: true - -resources: - containers: - - container: CentosContainer - image: mcr.microsoft.com/dotnet-buildtools/prereqs:centos-stream8-mlnet - - - container: UbuntuContainer - image: mcr.microsoft.com/dotnet-buildtools/prereqs:ubuntu-18.04-mlnet - -jobs: -- template: /build/ci/job-template.yml - parameters: - name: Centos_x64_Net80 - buildScript: ./build.sh - container: CentosContainer - customMatrixes: - Debug_Build: - _configuration: Debug - _config_short: DI - _targetFramework: net8.0 - _includeBenchmarkData: false - Release_Build: - _configuration: Release - _config_short: RI - _targetFramework: net8.0 - _includeBenchmarkData: true - nightlyBuild: true - pool: - vmImage: ubuntu-18.04 - -- template: /build/ci/job-template.yml - parameters: - name: Ubuntu_x64_NetCoreApp21 - buildScript: ./build.sh - container: UbuntuContainer - nightlyBuild: true - pool: - vmImage: ubuntu-18.04 - -- template: /build/ci/job-template.yml - parameters: - name: MacOS_x64_NetCoreApp21 - buildScript: ./build.sh - nightlyBuild: true - pool: - vmImage: macOS-12 - -- template: /build/ci/job-template.yml - parameters: - name: Windows_x64_Net80 - buildScript: build.cmd - customMatrixes: - Debug_Build: - _configuration: Debug - _config_short: DI - _includeBenchmarkData: false - Release_Build: - _configuration: Release - _config_short: RI - _includeBenchmarkData: true - nightlyBuild: true - pool: - vmImage: windows-2019 - -- template: /build/ci/job-template.yml - parameters: - name: Windows_x64_NetCoreApp21 - buildScript: build.cmd - nightlyBuild: true - pool: - vmImage: windows-2019 - -- template: /build/ci/job-template.yml - parameters: - name: Windows_x64_NetFx461 - buildScript: build.cmd - customMatrixes: - Debug_Build: - _configuration: Debug - _config_short: DFX - _includeBenchmarkData: false - _targetFramework: net462 - Release_Build: - _configuration: Release - _config_short: RFX - _includeBenchmarkData: false - _targetFramework: net462 - nightlyBuild: true - pool: - vmImage: windows-2019 - -- template: /build/ci/job-template.yml - parameters: - name: Windows_x86_NetCoreApp21 - architecture: x86 - buildScript: build.cmd - nightlyBuild: true - pool: - vmImage: windows-2019 diff --git a/build/.outer-loop-build.yml b/build/.outer-loop-build.yml deleted file mode 100644 index e96dab3bd4..0000000000 --- a/build/.outer-loop-build.yml +++ /dev/null @@ -1,111 +0,0 @@ -################################################################################ -# ML.NET's outer loop build that will run all tests -################################################################################ - -# no PR builds -pr: none - -# no CI builds -trigger: none - -# scheduled trigger, runs at UTC 8:00 every day which is midnight of GMT-8 -schedules: -- cron: "0 8 * * *" - displayName: outer loop build at midnight - branches: - include: - - main - - feature/* - - release/* - always: true - - -resources: - containers: - - container: CentosContainer - image: mcr.microsoft.com/dotnet-buildtools/prereqs:centos-stream8-mlnet - - - container: UbuntuContainer - image: mcr.microsoft.com/dotnet-buildtools/prereqs:ubuntu-18.04-mlnet - -jobs: -- template: /build/ci/job-template.yml - parameters: - name: Centos_x64_Net80 - buildScript: ./build.sh - container: CentosContainer - customMatrixes: - Debug_Build: - _configuration: Debug - _config_short: DI - _includeBenchmarkData: false - Release_Build: - _configuration: Release - _config_short: RI - _includeBenchmarkData: true - pool: - vmImage: ubuntu-18.04 - -- template: /build/ci/job-template.yml - parameters: - name: Ubuntu_x64_Net80 - buildScript: ./build.sh - container: UbuntuContainer - pool: - vmImage: ubuntu-18.04 - -- template: /build/ci/job-template.yml - parameters: - name: MacOS_x64_Net80 - buildScript: ./build.sh - pool: - vmImage: macOS-10.15 - -- template: /build/ci/job-template.yml - parameters: - name: Windows_x64_Net80 - buildScript: build.cmd - customMatrixes: - Debug_Build: - _configuration: Debug - _config_short: DI - _includeBenchmarkData: false - Release_Build: - _configuration: Release - _config_short: RI - _includeBenchmarkData: true - pool: - vmImage: windows-2019 - -- template: /build/ci/job-template.yml - parameters: - name: Windows_x64_Net80 - buildScript: build.cmd - pool: - vmImage: windows-2019 - -- template: /build/ci/job-template.yml - parameters: - name: Windows_x64_NetFx461 - buildScript: build.cmd - customMatrixes: - Debug_Build: - _configuration: Debug - _config_short: DFX - _includeBenchmarkData: false - _targetFramework: net462 - Release_Build: - _configuration: Release - _config_short: RFX - _includeBenchmarkData: false - _targetFramework: net462 - pool: - vmImage: windows-2019 - -- template: /build/ci/job-template.yml - parameters: - name: Windows_x86_Net80 - architecture: x86 - buildScript: build.cmd - pool: - vmImage: windows-2019 diff --git a/build/ci/job-template.yml b/build/ci/job-template.yml index 5b46ca04ae..8e458e076e 100644 --- a/build/ci/job-template.yml +++ b/build/ci/job-template.yml @@ -7,7 +7,6 @@ parameters: pool: {} customMatrixes: '' codeCoverage: false - nightlyBuild: false innerLoop: false runSpecific: false container: '' @@ -17,9 +16,7 @@ parameters: jobs: - job: ${{ parameters.name }} - ${{ if eq(parameters.nightlyBuild, 'true') }}: - timeoutInMinutes: 30 - ${{ if and(eq(parameters.nightlyBuild, 'false'), eq(parameters.codeCoverage, 'false')) }}: + ${{ if eq(parameters.codeCoverage, 'false') }}: timeoutInMinutes: 120 ${{ if eq(parameters.codeCoverage, 'true') }}: timeoutInMinutes: 150 @@ -27,9 +24,6 @@ jobs: variables: dotnetPath: $(Build.SourcesDirectory)/.dotnet/dotnet nugetFeed: https://pkgs.dev.azure.com/dnceng/public/_packaging/MachineLearning/nuget/v3/index.json - nightlyBuildProjPath: $(Build.SourcesDirectory)/test/Microsoft.ML.NightlyBuild.Tests/Microsoft.ML.NightlyBuild.Tests.csproj - nightlyBuildRunPath: $(Build.SourcesDirectory)/artifacts/bin/Microsoft.ML.NightlyBuild.Tests/$(_configuration)/$(_targetFramework) - runNightlyBuildProj: $(Build.SourcesDirectory)/test/run-night-build-tests.proj packageUpdaterProjPath: $(Build.SourcesDirectory)/test/Microsoft.ML.NugetPackageVersionUpdater/Microsoft.ML.NugetPackageVersionUpdater.csproj versionFilePath: $(Build.SourcesDirectory)/test/Microsoft.ML.NugetPackageVersionUpdater/latest_versions.txt PROCDUMP_PATH: '$(Build.SourcesDirectory)/Tools/ProcDump/' @@ -73,9 +67,6 @@ jobs: - ${{ if and(contains(parameters.pool.vmImage, 'macOS'), contains(parameters.name, 'cross')) }}: - script: brew update && brew install -f --overwrite python@3.12 && brew install libomp && brew link libomp --force displayName: Install MacOS ARM build dependencies - - ${{ if and( eq(parameters.nightlyBuild, 'true'), eq(parameters.pool.vmImage, 'ubuntu-18.04')) }}: - - bash: echo "##vso[task.setvariable variable=LD_LIBRARY_PATH]$(nightlyBuildRunPath):$LD_LIBRARY_PATH" - displayName: Set LD_LIBRARY_PATH for Ubuntu and CentOS to locate Native shared library in current running path - script: ${{ parameters.buildScript }} -configuration $(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} /p:Coverage=${{ parameters.codeCoverage }} $(testTargetFramework) /p:CreateHardLinksForCopyFilesToOutputDirectoryIfPossible=True $(hardLink) displayName: Build - ${{ if eq(parameters.pool.vmImage, 'macOS-10.15') }}: @@ -98,73 +89,50 @@ jobs: targetType: inline script: Get-ChildItem -Recurse -Directory | where {$_.FullName -like '*\runtimes\*'} | where {$_.FullName -notlike '*\artifacts\pkgassets*'} | where {$_.FullName -notlike '*\win*'} | sort length -Descending | Remove-Item -Recurse -Confirm:$false -Force -ErrorAction SilentlyContinue displayName: Clean up non-Windows runtime folders of NuGet Packages to save disk space - - ${{ if eq(parameters.nightlyBuild, 'true') }}: - - script: $(dotnetPath) restore $(nightlyBuildProjPath) - displayName: Restore nightly build project - - script: $(dotnetPath) list $(nightlyBuildProjPath) package --source $(nugetFeed) --outdated > $(versionFilePath) - displayName: List latest package versions - - script: $(dotnetPath) run --project $(packageUpdaterProjPath) - displayName: Update package versions for nightly build - - ${{ if eq(parameters.buildScript, 'build.cmd') }}: - - powershell: | - Get-ChildItem -Path '.\artifacts\bin\*' -Recurse | - Select -ExpandProperty FullName | - Where {$_ -NotMatch '.*\\Microsoft\.ML\.NightlyBuild\.Tests.*|.*\\Native.*'} | - sort length -Descending | - Remove-Item -force - Write-Output "Done cleaning up usless project..." - displayName: Clean up useless project - - script: $(dotnetPath) msbuild -restore $(nightlyBuildProjPath) /p:ReferenceTypeForTestFramework="Nuget" /p:Configuration=$(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} $(testTargetFramework) - displayName: Build Nightly-Build Project with latest package versions - - script: $(dotnetPath) msbuild $(runNightlyBuildProj) /t:RunNightlyBuildTests /p:Configuration=$(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} $(testTargetFramework) - displayName: Run Nightly Build Tests - - ${{ if eq(parameters.nightlyBuild, 'false') }}: - - ${{ if eq(parameters.innerLoop, 'false') }}: - - ${{ if and(eq(parameters.runSpecific, 'false'), eq(parameters.useVSTestTask, 'false')) }}: - - script: set PATH=%PATH%;%USERPROFILE%\.nuget\packages\libtorch-cpu-win-x64\2.1.0.1\runtimes\win-x64\native;%USERPROFILE%\.nuget\packages\torchsharp\0.101.5\runtimes\win-x64\native & ${{ parameters.buildScript }} /p:Build=false -configuration $(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} -test -integrationTest /p:Coverage=${{ parameters.codeCoverage }} $(testTargetFramework) - displayName: Run All Tests. - - ${{ if and(eq(parameters.runSpecific, 'true'), eq(parameters.useVSTestTask, 'false')) }}: - - script: ${{ parameters.buildScript }} /p:Build=false -configuration $(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} -test -integrationTest /p:TestRunnerAdditionalArguments='-trait$(spaceValue)Category=RunSpecificTest' /p:Coverage=${{ parameters.codeCoverage }} $(testTargetFramework) - displayName: Run Specific Tests. - - ${{ if and(eq(parameters.buildScript, 'build.cmd'), eq(parameters.useVSTestTask, 'true')) }}: - - task: VSTest@2 - displayName: Run Tests with VSTest - inputs: - testSelector: testAssemblies - testAssemblyVer2: | - **\*test.dll - **\*tests.dll - !**\obj\** - runSettingsFile: $(Build.SourcesDirectory)/tools-local/vstest.runsettings - searchFolder: '$(System.DefaultWorkingDirectory)' - vstestLocationMethod: 'version' - vsTestVersion: 'latest' - runInParallel: False - runTestsInIsolation: True - codeCoverageEnabled: ${{ parameters.codeCoverage }} - otherConsoleOptions: ${{ parameters.vsTestConfiguration }} - dontDistribute: True - diagnosticsEnabled: False - collectDumpOn: onAbortOnly - publishRunAttachments: true - - ${{ if eq(parameters.innerLoop, 'true') }}: - - ${{ if eq(parameters.codeCoverage, True) }}: - - script: ${{ parameters.buildScript }} /p:Build=false -configuration $(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} -test -integrationTest /p:TestRunnerAdditionalArguments='-notrait$(spaceValue)Category=SkipInCI' /p:Coverage=${{ parameters.codeCoverage }} $(testTargetFramework) - displayName: Run CI Tests. - - ${{ if eq(parameters.codeCoverage, False) }}: - - template: /build/ci/send-to-helix.yml - parameters: - HelixTargetQueues: ${{ parameters.helixQueue }} - Configuration: $(_configuration) - Architecture: ${{ parameters.architecture }} - ${{ if eq(parameters.buildScript, 'build.cmd') }}: - MsBuildScript: 'powershell $(Build.SourcesDirectory)/eng/common/msbuild.ps1' - WarnAsError: '-warnAsError 0' - TestTargetFramework: '$(testTargetFramework)' - ${{ if eq(parameters.buildScript, './build.sh') }}: - MsBuildScript: '$(Build.SourcesDirectory)/eng/common/msbuild.sh' - WarnAsError: '--warnAsError false' - TestTargetFramework: '$(testTargetFramework)' + + - ${{ if eq(parameters.innerLoop, 'false') }}: + - ${{ if and(eq(parameters.runSpecific, 'false'), eq(parameters.useVSTestTask, 'false')) }}: + - script: set PATH=%PATH%;%USERPROFILE%\.nuget\packages\libtorch-cpu-win-x64\2.1.0.1\runtimes\win-x64\native;%USERPROFILE%\.nuget\packages\torchsharp\0.101.5\runtimes\win-x64\native & ${{ parameters.buildScript }} /p:Build=false -configuration $(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} -test -integrationTest /p:Coverage=${{ parameters.codeCoverage }} + displayName: Run All Tests. + - ${{ if and(eq(parameters.runSpecific, 'true'), eq(parameters.useVSTestTask, 'false')) }}: + - script: ${{ parameters.buildScript }} /p:Build=false -configuration $(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} -test -integrationTest /p:TestRunnerAdditionalArguments='-trait$(spaceValue)Category=RunSpecificTest' /p:Coverage=${{ parameters.codeCoverage }} + displayName: Run Specific Tests. + - ${{ if and(eq(parameters.buildScript, 'build.cmd'), eq(parameters.useVSTestTask, 'true')) }}: + - task: VSTest@2 + displayName: Run Tests with VSTest + inputs: + testSelector: testAssemblies + testAssemblyVer2: | + **\*test.dll + **\*tests.dll + !**\obj\** + runSettingsFile: $(Build.SourcesDirectory)/tools-local/vstest.runsettings + searchFolder: '$(System.DefaultWorkingDirectory)' + vstestLocationMethod: 'version' + vsTestVersion: 'latest' + runInParallel: False + runTestsInIsolation: True + codeCoverageEnabled: ${{ parameters.codeCoverage }} + dontDistribute: True + diagnosticsEnabled: False + collectDumpOn: onAbortOnly + publishRunAttachments: true + - ${{ if eq(parameters.innerLoop, 'true') }}: + - ${{ if eq(parameters.codeCoverage, True) }}: + - script: ${{ parameters.buildScript }} /p:Build=false -configuration $(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} -test -integrationTest /p:TestRunnerAdditionalArguments='-notrait$(spaceValue)Category=SkipInCI' /p:Coverage=${{ parameters.codeCoverage }} + displayName: Run CI Tests. + - ${{ if eq(parameters.codeCoverage, False) }}: + - template: /build/ci/send-to-helix.yml + parameters: + HelixTargetQueues: ${{ parameters.helixQueue }} + Configuration: $(_configuration) + Architecture: ${{ parameters.architecture }} + ${{ if eq(parameters.buildScript, 'build.cmd') }}: + MsBuildScript: 'powershell $(Build.SourcesDirectory)/eng/common/msbuild.ps1' + WarnAsError: '-warnAsError 0' + ${{ if eq(parameters.buildScript, './build.sh') }}: + MsBuildScript: '$(Build.SourcesDirectory)/eng/common/msbuild.sh' + WarnAsError: '--warnAsError false' - script: $(dotnetPath) msbuild -restore build/Codecoverage.proj displayName: Upload coverage to codecov.io condition: and(succeeded(), eq(${{ parameters.codeCoverage }}, True)) @@ -235,6 +203,5 @@ jobs: targetType: inline script: Get-ChildItem -Path '.\artifacts\bin' -Recurse | Where-Object {$_.FullName -like "*runtimes*"} | Remove-Item -Recurse -Confirm:$false -Force displayName: Clean up runtime folder for package (Unix) - - ${{ if eq(parameters.nightlyBuild, 'false') }}: - - script: ${{ parameters.buildScript }} /p:Build=false -pack -ci -configuration $(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} $(testTargetFramework) - displayName: Build Packages + - script: ${{ parameters.buildScript }} /p:Build=false -pack -ci -configuration $(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} + displayName: Build Packages From c5718cb08125008cfed8c27a8c3bd0e6d3d913ee Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Fri, 22 Mar 2024 16:21:44 -0700 Subject: [PATCH 09/26] Refactor helix testing to get test info from projects Rather than constructing the test information outside the projects make a target within the test reponsible for this. That allows us to avoid hardcoding paths, and directly control the behavior for the individual test at the project level. --- .vsts-dotnet-ci.yml | 51 ++--------- build/ci/job-template.yml | 6 +- build/ci/send-to-helix.yml | 3 - build/codecoverage-ci.yml | 1 - eng/helix.proj | 75 ++--------------- test/Directory.Build.props | 6 +- test/Directory.Build.targets | 84 +++++++++++++++++++ ...oft.Data.Analysis.Interactive.Tests.csproj | 2 +- ...soft.Data.Analysis.PerformanceTests.csproj | 2 +- .../Microsoft.ML.Benchmarks.Tests.csproj | 4 + ...crosoft.ML.CpuMath.PerformanceTests.csproj | 3 +- .../Microsoft.ML.NightlyBuild.Tests.csproj | 1 + ...osoft.ML.NugetPackageVersionUpdater.csproj | 1 + .../Microsoft.ML.PerformanceTests.csproj | 1 + .../Microsoft.ML.TensorFlow.Tests.csproj | 1 + .../Microsoft.ML.TestFrameworkCommon.csproj | 4 + .../Microsoft.ML.TorchSharp.Tests.csproj | 1 + 17 files changed, 120 insertions(+), 126 deletions(-) diff --git a/.vsts-dotnet-ci.yml b/.vsts-dotnet-ci.yml index 485b8fb2f1..2bef29dc49 100644 --- a/.vsts-dotnet-ci.yml +++ b/.vsts-dotnet-ci.yml @@ -43,7 +43,7 @@ jobs: - template: /build/ci/job-template.yml parameters: architecture: arm - name: Ubuntu_x64_cross_arm_Net80 + name: Ubuntu_x64_cross_arm buildScript: ./build.sh container: UbuntuCrossArmContainer customMatrixes: @@ -51,12 +51,10 @@ jobs: _configuration: Debug _config_short: DI _includeBenchmarkData: false - _targetFramework: net8.0 Release_Build: _configuration: Release _config_short: RI _includeBenchmarkData: true - _targetFramework: net8.0 innerLoop: true pool: name: NetCore-Public @@ -66,7 +64,7 @@ jobs: - template: /build/ci/job-template.yml parameters: architecture: arm64 - name: Ubuntu_x64_cross_arm64_Net80 + name: Ubuntu_x64_cross_arm64 buildScript: ./build.sh container: UbuntuCrossArm64Container customMatrixes: @@ -74,12 +72,10 @@ jobs: _configuration: Debug _config_short: DI _includeBenchmarkData: false - _targetFramework: net8.0 Release_Build: _configuration: Release _config_short: RI _includeBenchmarkData: true - _targetFramework: net8.0 innerLoop: true pool: name: NetCore-Public @@ -88,7 +84,7 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: Centos_x64_Net80 + name: Centos_x64 buildScript: ./build.sh container: CentosContainer innerLoop: true @@ -99,7 +95,7 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: Ubuntu_x64_Net80 + name: Ubuntu_x64 buildScript: ./build.sh container: UbuntuContainer innerLoop: true @@ -110,7 +106,7 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: MacOS_x64_Net80 + name: MacOS_x64 buildScript: ./build.sh innerLoop: true pool: @@ -120,19 +116,17 @@ jobs: - template: /build/ci/job-template.yml parameters: architecture: arm64 - name: MacOS_cross_arm64_Net6 + name: MacOS_cross_arm64 buildScript: ./build.sh customMatrixes: Debug_Build: _configuration: Debug _config_short: DI _includeBenchmarkData: false - _targetFramework: net8.0 Release_Build: _configuration: Release _config_short: RI _includeBenchmarkData: true - _targetFramework: net8.0 innerLoop: true pool: vmImage: macOS-12 @@ -141,21 +135,18 @@ jobs: - template: /build/ci/job-template.yml parameters: architecture: arm64 - name: Windows_cross_arm64_Net6 + name: Windows_cross_arm64 buildScript: build.cmd customMatrixes: Debug_Build: _configuration: Debug _config_short: DI _includeBenchmarkData: false - _targetFramework: net8.0 Release_Build: _configuration: Release _config_short: RI _includeBenchmarkData: false - _targetFramework: net8.0 innerLoop: true - vsTestConfiguration: "/Framework:.NETCoreApp,Version=v6.0" pool: name: NetCore-Public demands: ImageOverride -equals 1es-windows-2019-open @@ -163,10 +154,9 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: Windows_x64_Net6 + name: Windows_x64 buildScript: build.cmd innerLoop: true - vsTestConfiguration: "/Framework:.NETCoreApp,Version=v3.1" pool: name: NetCore-Public demands: ImageOverride -equals 1es-windows-2019-open @@ -174,33 +164,10 @@ jobs: - template: /build/ci/job-template.yml parameters: - name: Windows_x64_NetFx461 - buildScript: build.cmd - customMatrixes: - Debug_Build: - _configuration: Debug - _config_short: DFX - _includeBenchmarkData: false - _targetFramework: net462 - Release_Build: - _configuration: Release - _config_short: RFX - _includeBenchmarkData: false - _targetFramework: net462 - innerLoop: true - vsTestConfiguration: "/Framework:.NETCoreApp,Version=v4.0" - pool: - name: NetCore-Public - demands: ImageOverride -equals 1es-windows-2019-open - helixQueue: Windows.10.Amd64.Open - -- template: /build/ci/job-template.yml - parameters: - name: Windows_x86_Net6 + name: Windows_x86 architecture: x86 buildScript: build.cmd innerLoop: true - vsTestConfiguration: "/Framework:.NETCoreApp,Version=v3.1" pool: name: NetCore-Public demands: ImageOverride -equals 1es-windows-2019-open diff --git a/build/ci/job-template.yml b/build/ci/job-template.yml index 8e458e076e..171d61297e 100644 --- a/build/ci/job-template.yml +++ b/build/ci/job-template.yml @@ -12,7 +12,6 @@ parameters: container: '' useVSTestTask: false helixQueue: '' - testTargetFramework: '' jobs: - job: ${{ parameters.name }} @@ -37,7 +36,6 @@ jobs: hardLink: '' ${{ if eq(parameters.codeCoverage, 'false') }}: hardLink: '/p:CreateHardLinksForCopyLocalIfPossible=True' - testTargetFramework: '/p:TestTargetFramework=$(_targetFramework)' strategy: matrix: ${{ if eq(parameters.customMatrixes, '') }}: @@ -45,12 +43,10 @@ jobs: _configuration: Debug _config_short: D _includeBenchmarkData: false - _targetFramework: net8.0 Release_Build: _configuration: Release _config_short: R _includeBenchmarkData: true - _targetFramework: net8.0 ${{ if ne(parameters.customMatrixes, '') }}: ${{ insert }}: ${{ parameters.customMatrixes }} @@ -67,7 +63,7 @@ jobs: - ${{ if and(contains(parameters.pool.vmImage, 'macOS'), contains(parameters.name, 'cross')) }}: - script: brew update && brew install -f --overwrite python@3.12 && brew install libomp && brew link libomp --force displayName: Install MacOS ARM build dependencies - - script: ${{ parameters.buildScript }} -configuration $(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} /p:Coverage=${{ parameters.codeCoverage }} $(testTargetFramework) /p:CreateHardLinksForCopyFilesToOutputDirectoryIfPossible=True $(hardLink) + - script: ${{ parameters.buildScript }} -configuration $(_configuration) /p:TargetArchitecture=${{ parameters.architecture }} /p:TestArchitectures=${{ parameters.architecture }} /p:Coverage=${{ parameters.codeCoverage }} /p:CreateHardLinksForCopyFilesToOutputDirectoryIfPossible=True $(hardLink) displayName: Build - ${{ if eq(parameters.pool.vmImage, 'macOS-10.15') }}: - task: Bash@3 diff --git a/build/ci/send-to-helix.yml b/build/ci/send-to-helix.yml index 9042bc62e7..2023299e07 100644 --- a/build/ci/send-to-helix.yml +++ b/build/ci/send-to-helix.yml @@ -9,7 +9,6 @@ parameters: Architecture: 'x64' MsBuildScript: '' WarnAsError: '' - TestTargetFramework: '' HelixConfiguration: '' # optional -- additional property attached to a job EnableXUnitReporter: true # optional -- true enables XUnit result reporting to Mission Control WaitForWorkItemCompletion: true # optional -- true will make the task wait until work items have been completed and fail the build if work items fail. False is "fire and forget." @@ -27,7 +26,6 @@ steps: /p:Configuration=${{ parameters.Configuration }} /p:TargetArchitecture=${{ parameters.Architecture }} /p:BuildConfig=${{ parameters.Configuration }} - /p:BuildArchitecture=${{ parameters.Architecture }} /p:HelixSource=${{ parameters.HelixSource }} /p:HelixType=${{ parameters.HelixType }} /p:HelixBuild=${{ parameters.HelixBuild }} @@ -37,7 +35,6 @@ steps: /p:WaitForWorkItemCompletion=${{ parameters.WaitForWorkItemCompletion }} /p:HelixBaseUri=${{ parameters.HelixBaseUri }} ${{ parameters.WarnAsError }} - ${{ parameters.TestTargetFramework }} displayName: ${{ parameters.DisplayNamePrefix }} env: SYSTEM_ACCESSTOKEN: $(System.AccessToken) diff --git a/build/codecoverage-ci.yml b/build/codecoverage-ci.yml index 2f610b6c0e..b7daa2e3cd 100644 --- a/build/codecoverage-ci.yml +++ b/build/codecoverage-ci.yml @@ -35,7 +35,6 @@ jobs: _configuration: Debug _config_short: D _includeBenchmarkData: false - _targetFramework: net8.0 codeCoverage: true pool: name: NetCore-Public diff --git a/eng/helix.proj b/eng/helix.proj index 7a3dc0f17e..d3b8cc2525 100644 --- a/eng/helix.proj +++ b/eng/helix.proj @@ -30,9 +30,9 @@ $([System.Text.RegularExpressions.Regex]::Match($(GlobalJsonContent), '(%3F<="dotnet": ").*(%3F=")')) - win-x86 - linux-arm - osx-arm64 + win-x86 + linux-arm + osx-arm64 true @@ -41,8 +41,8 @@ runtime - win-x86 - linux-arm + win-x86 + linux-arm @@ -63,15 +63,6 @@ - - - - - - - - - @@ -91,10 +82,10 @@ - + @@ -107,58 +98,6 @@ $(HelixPreCommands);export LD_LIBRARY_PATH=/opt/homebrew/opt/mono-libgdiplus/lib;ls /usr/lib;ls $HELIX_WORKITEM_ROOT $(HelixPreCommands);sudo apt update;sudo apt-get install libomp-dev libomp5 -y - - $HELIX_CORRELATION_PAYLOAD - %HELIX_CORRELATION_PAYLOAD% - - /usr/local/opt/libomp/lib/libiomp5.dylib;/usr/local/opt/libomp/lib/libomp.dylib; - /usr/local/opt/libomp/lib/libomp.dylib; - - - - - - - - - - - - - - - - - - - - - - - - - - - $(BUILD_SOURCESDIRECTORY)\artifacts\bin\%(ProjectsWithTargetFramework.Filename)\$(BuildConfig)\%(ProjectsWithTargetFramework.TargetFrameworks)\ - $(BUILD_SOURCESDIRECTORY)\artifacts\bin\%(ProjectsWithTargetFramework.Filename)\$(BuildConfig)\%(ProjectsWithTargetFramework.TargetFrameworks)\win-x64 - dotnet exec --roll-forward Major --runtimeconfig %(ProjectsWithTargetFramework.Filename).runtimeconfig.json --depsfile %(ProjectsWithTargetFramework.Filename).deps.json $(HelixCorrelationPayloadPath)/xunit-runner/tools/netcoreapp2.0/xunit.console.dll %(ProjectsWithTargetFramework.Filename).dll -notrait Category=SkipInCI -xml testResults.xml - $(HelixCorrelationPayloadPath)/xunit-runner/tools/net462/xunit.console.exe %(ProjectsWithTargetFramework.Filename).dll -notrait Category=SkipInCI -xml testResults.xml - 01:00:00 - 00:30:00 - $(WorkItemTimeout) - - diff --git a/test/Directory.Build.props b/test/Directory.Build.props index 980a2dbc35..fa02e6e1c4 100644 --- a/test/Directory.Build.props +++ b/test/Directory.Build.props @@ -2,10 +2,8 @@ - $(NetMinimum);$(NetCurrent) - $(TargetFrameworks);$(NetFrameworkMinimum) - $(TestTargetFramework) - win-x64 + $(NetMinimum);$(NetCurrent);$(NetFrameworkMinimum) + win-x64 true false true diff --git a/test/Directory.Build.targets b/test/Directory.Build.targets index 3030dc9156..ed7c32c61a 100644 --- a/test/Directory.Build.targets +++ b/test/Directory.Build.targets @@ -11,4 +11,88 @@ + + + true + + true + + $HELIX_CORRELATION_PAYLOAD + %HELIX_CORRELATION_PAYLOAD% + + + + _GetHelixWorkItemsOuter + _GetHelixWorkItemsInner + + + + + + + + + + + + + + $(TargetDir) + dotnet exec --roll-forward Major --runtimeconfig $(TargetName).runtimeconfig.json --depsfile $(TargetName).deps.json $(HelixCorrelationPayloadPath)/xunit-runner/tools/netcoreapp2.0/xunit.console.dll $(TargetName)).dll -notrait Category=SkipInCI -xml testResults.xml + $(HelixCorrelationPayloadPath)/xunit-runner/tools/net462/xunit.console.exe $(TargetName).dll -notrait Category=SkipInCI -xml testResults.xml + 01:00:00 + 00:30:00 + $(WorkItemTimeout) + + + + + + + /usr/local/opt/libomp/lib/libiomp5.dylib;/usr/local/opt/libomp/lib/libomp.dylib; + /usr/local/opt/libomp/lib/libomp.dylib; + + + + + + + + + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj b/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj index f9b4850bd1..44f6d29bc7 100644 --- a/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj +++ b/test/Microsoft.Data.Analysis.Interactive.Tests/Microsoft.Data.Analysis.Interactive.Tests.csproj @@ -1,7 +1,7 @@  - $(NetMinimum);$(NetCurrent) + $(NetMinimum);$(NetCurrent) $(NoWarn);MSML_ExtendBaseTestClass diff --git a/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj b/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj index dadf551d52..39e6f38efa 100644 --- a/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj +++ b/test/Microsoft.Data.Analysis.PerformanceTests/Microsoft.Data.Analysis.PerformanceTests.csproj @@ -3,7 +3,7 @@ Exe disable - $(NetMinimum);$(NetCurrent) + $(NetMinimum);$(NetCurrent) false diff --git a/test/Microsoft.ML.Benchmarks.Tests/Microsoft.ML.Benchmarks.Tests.csproj b/test/Microsoft.ML.Benchmarks.Tests/Microsoft.ML.Benchmarks.Tests.csproj index 26fc578874..f619da6ef4 100644 --- a/test/Microsoft.ML.Benchmarks.Tests/Microsoft.ML.Benchmarks.Tests.csproj +++ b/test/Microsoft.ML.Benchmarks.Tests/Microsoft.ML.Benchmarks.Tests.csproj @@ -1,5 +1,9 @@  + + true + + diff --git a/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj b/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj index 069a75d4ad..587421a441 100644 --- a/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj +++ b/test/Microsoft.ML.CpuMath.PerformanceTests/Microsoft.ML.CpuMath.PerformanceTests.csproj @@ -3,10 +3,11 @@ Exe false - $(NetMinimum);$(NetCurrent) + $(NetMinimum);$(NetCurrent) true false true + true diff --git a/test/Microsoft.ML.NightlyBuild.Tests/Microsoft.ML.NightlyBuild.Tests.csproj b/test/Microsoft.ML.NightlyBuild.Tests/Microsoft.ML.NightlyBuild.Tests.csproj index 3997e12a2d..7d4002a516 100644 --- a/test/Microsoft.ML.NightlyBuild.Tests/Microsoft.ML.NightlyBuild.Tests.csproj +++ b/test/Microsoft.ML.NightlyBuild.Tests/Microsoft.ML.NightlyBuild.Tests.csproj @@ -6,6 +6,7 @@ false false + true diff --git a/test/Microsoft.ML.NugetPackageVersionUpdater/Microsoft.ML.NugetPackageVersionUpdater.csproj b/test/Microsoft.ML.NugetPackageVersionUpdater/Microsoft.ML.NugetPackageVersionUpdater.csproj index 0b8b27ab9e..1fe2db4b92 100644 --- a/test/Microsoft.ML.NugetPackageVersionUpdater/Microsoft.ML.NugetPackageVersionUpdater.csproj +++ b/test/Microsoft.ML.NugetPackageVersionUpdater/Microsoft.ML.NugetPackageVersionUpdater.csproj @@ -4,6 +4,7 @@ Exe Microsoft.ML.NugetPackageVersionUpdater.Program $(MSBuildProjectDirectory) + true diff --git a/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj b/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj index 55b9b7c6d4..9a625eb8fb 100644 --- a/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj +++ b/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj @@ -5,6 +5,7 @@ false false true + true diff --git a/test/Microsoft.ML.TensorFlow.Tests/Microsoft.ML.TensorFlow.Tests.csproj b/test/Microsoft.ML.TensorFlow.Tests/Microsoft.ML.TensorFlow.Tests.csproj index 78085851c2..57cd12ad52 100644 --- a/test/Microsoft.ML.TensorFlow.Tests/Microsoft.ML.TensorFlow.Tests.csproj +++ b/test/Microsoft.ML.TensorFlow.Tests/Microsoft.ML.TensorFlow.Tests.csproj @@ -4,6 +4,7 @@ Microsoft.ML.TensorFlow.Tests Test + true diff --git a/test/Microsoft.ML.TestFrameworkCommon/Microsoft.ML.TestFrameworkCommon.csproj b/test/Microsoft.ML.TestFrameworkCommon/Microsoft.ML.TestFrameworkCommon.csproj index cd0c6cc861..8492d60fe0 100644 --- a/test/Microsoft.ML.TestFrameworkCommon/Microsoft.ML.TestFrameworkCommon.csproj +++ b/test/Microsoft.ML.TestFrameworkCommon/Microsoft.ML.TestFrameworkCommon.csproj @@ -1,5 +1,9 @@  + + true + + diff --git a/test/Microsoft.ML.TorchSharp.Tests/Microsoft.ML.TorchSharp.Tests.csproj b/test/Microsoft.ML.TorchSharp.Tests/Microsoft.ML.TorchSharp.Tests.csproj index 4d7de3e37e..5a8c73901e 100644 --- a/test/Microsoft.ML.TorchSharp.Tests/Microsoft.ML.TorchSharp.Tests.csproj +++ b/test/Microsoft.ML.TorchSharp.Tests/Microsoft.ML.TorchSharp.Tests.csproj @@ -6,6 +6,7 @@ None + true From d26b856d2cb25797780ba84104e8508a8eb06d22 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Mon, 25 Mar 2024 16:13:47 -0700 Subject: [PATCH 10/26] Fix mac typo & only build netfx on Windows --- test/Directory.Build.props | 1 + test/Directory.Build.targets | 2 +- 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/test/Directory.Build.props b/test/Directory.Build.props index fa02e6e1c4..2a0c98e121 100644 --- a/test/Directory.Build.props +++ b/test/Directory.Build.props @@ -3,6 +3,7 @@ $(NetMinimum);$(NetCurrent);$(NetFrameworkMinimum) + $(TargetFrameworks);$(NetFrameworkMinimum) win-x64 true false diff --git a/test/Directory.Build.targets b/test/Directory.Build.targets index ed7c32c61a..9d79f2da00 100644 --- a/test/Directory.Build.targets +++ b/test/Directory.Build.targets @@ -73,7 +73,7 @@ Command="install_name_tool -change "/usr/local/opt/libomp/lib/libomp.dylib" "@loader_path/libomp.dylib" $(TargetDir)libSymSgdNative.dylib" /> + Command="install_name_tool -change "/usr/local/opt/libomp/lib/libomp.dylib" "@loader_path/libomp.dylib" $(TargetDir)runtimes/osx-x64/native/lib_lightgbm.dylib" /> From dc0e5098d0c96a113b03c9128fe986b610569d42 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Mon, 25 Mar 2024 18:03:32 -0700 Subject: [PATCH 11/26] Update Directory.Build.targets --- test/Directory.Build.targets | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/test/Directory.Build.targets b/test/Directory.Build.targets index 9d79f2da00..6f4b2fc85e 100644 --- a/test/Directory.Build.targets +++ b/test/Directory.Build.targets @@ -45,7 +45,7 @@ $(TargetDir) - dotnet exec --roll-forward Major --runtimeconfig $(TargetName).runtimeconfig.json --depsfile $(TargetName).deps.json $(HelixCorrelationPayloadPath)/xunit-runner/tools/netcoreapp2.0/xunit.console.dll $(TargetName)).dll -notrait Category=SkipInCI -xml testResults.xml + dotnet exec --roll-forward Major --runtimeconfig $(TargetName).runtimeconfig.json --depsfile $(TargetName).deps.json $(HelixCorrelationPayloadPath)/xunit-runner/tools/netcoreapp2.0/xunit.console.dll $(TargetName).dll -notrait Category=SkipInCI -xml testResults.xml $(HelixCorrelationPayloadPath)/xunit-runner/tools/net462/xunit.console.exe $(TargetName).dll -notrait Category=SkipInCI -xml testResults.xml 01:00:00 00:30:00 @@ -95,4 +95,4 @@ Files="@(WindowsFiles);@(OsxFiles)" /> - \ No newline at end of file + From 234f41a346c1b4e261f9396c5b13ea17400b931c Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Mon, 25 Mar 2024 18:27:39 -0700 Subject: [PATCH 12/26] Update Directory.Build.props --- test/Directory.Build.props | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/test/Directory.Build.props b/test/Directory.Build.props index 2a0c98e121..575d40302a 100644 --- a/test/Directory.Build.props +++ b/test/Directory.Build.props @@ -2,7 +2,7 @@ - $(NetMinimum);$(NetCurrent);$(NetFrameworkMinimum) + $(NetMinimum);$(NetCurrent) $(TargetFrameworks);$(NetFrameworkMinimum) win-x64 true From 8fb1f4fb271638ae22d452d4eea2289c7e53bdbc Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Tue, 26 Mar 2024 12:21:58 -0700 Subject: [PATCH 13/26] Separate netfx test file for core_manifest.json --- .../{ => netcoreapp}/core_manifest.json | 0 test/BaselineOutput/Common/core_manifest.json | 30768 ++++++++++++++++ 2 files changed, 30768 insertions(+) rename test/BaselineOutput/Common/EntryPoints/{ => netcoreapp}/core_manifest.json (100%) create mode 100644 test/BaselineOutput/Common/core_manifest.json diff --git a/test/BaselineOutput/Common/EntryPoints/core_manifest.json b/test/BaselineOutput/Common/EntryPoints/netcoreapp/core_manifest.json similarity index 100% rename from test/BaselineOutput/Common/EntryPoints/core_manifest.json rename to test/BaselineOutput/Common/EntryPoints/netcoreapp/core_manifest.json diff --git a/test/BaselineOutput/Common/core_manifest.json b/test/BaselineOutput/Common/core_manifest.json new file mode 100644 index 0000000000..3dbaad5989 --- /dev/null +++ b/test/BaselineOutput/Common/core_manifest.json @@ -0,0 +1,30768 @@ +{ + "EntryPoints": [ + { + "Name": "Data.CustomTextLoader", + "Desc": "Import a dataset from a text file", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "InputFile", + "Type": "FileHandle", + "Desc": "Location of the input file", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "CustomSchema", + "Type": "String", + "Desc": "Custom schema to use for parsing", + "Aliases": [ + "schema" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The resulting data view" + } + ] + }, + { + "Name": "Data.DataViewReference", + "Desc": "Pass dataview from memory to experiment", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Pointer to IDataView in memory", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The resulting data view" + } + ] + }, + { + "Name": "Data.IDataViewArrayConverter", + "Desc": "Create an array variable of IDataView", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "The data sets", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "The data set array" + } + ] + }, + { + "Name": "Data.PredictorModelArrayConverter", + "Desc": "Create an array variable of PredictorModel", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputModels", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The model array" + } + ] + }, + { + "Name": "Data.TextLoader", + "Desc": "Import a dataset from a text file", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "InputFile", + "Type": "FileHandle", + "Desc": "Location of the input file", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Arguments", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the column", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Type", + "Type": { + "Kind": "Enum", + "Values": [ + "I1", + "U1", + "I2", + "U2", + "I4", + "U4", + "I8", + "U8", + "R4", + "Num", + "R8", + "TX", + "Text", + "TXT", + "BL", + "Bool", + "TimeSpan", + "TS", + "DT", + "DateTime", + "DZ", + "DateTimeZone", + "UG", + "U16" + ] + }, + "Desc": "Type of the items in the column", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "R4" + }, + { + "Name": "Source", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Min", + "Type": "Int", + "Desc": "First index in the range", + "Required": true, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Max", + "Type": "Int", + "Desc": "Last index in the range", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "AutoEnd", + "Type": "Bool", + "Desc": "This range extends to the end of the line, but should be a fixed number of items", + "Aliases": [ + "auto" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "VariableEnd", + "Type": "Bool", + "Desc": "This range extends to the end of the line, which can vary from line to line", + "Aliases": [ + "var" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "AllOther", + "Type": "Bool", + "Desc": "This range includes only other indices not specified", + "Aliases": [ + "other" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "ForceVector", + "Type": "Bool", + "Desc": "Force scalar columns to be treated as vectors of length one", + "Aliases": [ + "vector" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + } + ] + } + }, + "Desc": "Source index range(s) of the column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "KeyCount", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Count", + "Type": "UInt", + "Desc": "Count of valid key values", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ] + }, + "Desc": "For a key column, this defines the range of values", + "Aliases": [ + "key" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "Column groups. Each group is specified as name:type:numeric-ranges, eg, col=Features:R4:1-17,26,35-40", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "AllowQuoting", + "Type": "Bool", + "Desc": "Whether the input may include double-quoted values. This parameter is used to distinguish separator characters in an input value from actual separators. When true, separators within double quotes are treated as part of the input value. When false, all separators, even those within quotes, are treated as delimiting a new column.", + "Aliases": [ + "quote" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "AllowSparse", + "Type": "Bool", + "Desc": "Whether the input may include sparse representations", + "Aliases": [ + "sparse" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "InputSize", + "Type": "Int", + "Desc": "Number of source columns in the text data. Default is that sparse rows contain their size information.", + "Aliases": [ + "size" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Separator", + "Type": { + "Kind": "Array", + "ItemType": "Char" + }, + "Desc": "Source column separator.", + "Aliases": [ + "sep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + "\t" + ] + }, + { + "Name": "Decimal Marker", + "Type": "Char", + "Desc": "Character symbol used to separate the integer part from the fractional part of a number written in decimal form.", + "Aliases": [ + "decimal" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "." + }, + { + "Name": "TrimWhitespace", + "Type": "Bool", + "Desc": "Remove trailing whitespace from lines", + "Aliases": [ + "trim" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "HasHeader", + "Type": "Bool", + "Desc": "Data file has header with feature names. Header is read only if options 'hs' and 'hf' are not specified.", + "Aliases": [ + "header" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseThreads", + "Type": "Bool", + "Desc": "Use separate parsing threads?", + "Aliases": [ + "threads" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "ReadMultilines", + "Type": "Bool", + "Desc": "Escape new line characters inside a quoted field? If AllowQuoting is false, this argument is ignored.", + "Aliases": [ + "multilines" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "HeaderFile", + "Type": "String", + "Desc": "File containing a header with feature names. If specified, header defined in the data file (header+) is ignored.", + "Aliases": [ + "hf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "MaxRows", + "Type": "Int", + "Desc": "Maximum number of rows to produce", + "Aliases": [ + "rows" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "EscapeChar", + "Type": "Char", + "Desc": "Character to use to escape quotes inside quoted fields. It can't be a character used as separator.", + "Aliases": [ + "escapechar" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "\"" + }, + { + "Name": "MissingRealsAsNaNs", + "Type": "Bool", + "Desc": "If true, empty float fields will be loaded as NaN. If false, they'll be loaded as 0. Default is false.", + "Aliases": [ + "missingrealnan" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + } + ] + }, + "Desc": "Arguments", + "Aliases": [ + "args" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The resulting data view" + } + ] + }, + { + "Name": "Models.AnomalyDetectionEvaluator", + "Desc": "Evaluates an anomaly detection scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "K", + "Type": "Int", + "Desc": "Expected number of false positives", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "P", + "Type": "Float", + "Desc": "Expected false positive rate", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01 + }, + { + "Name": "NumTopResults", + "Type": "Int", + "Desc": "Number of top-scored predictions to display", + "Aliases": [ + "n" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 50 + }, + { + "Name": "Stream", + "Type": "Bool", + "Desc": "Whether to calculate metrics in one pass", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxAucExamples", + "Type": "Int", + "Desc": "The number of samples to use for AUC calculation. If 0, AUC is not computed. If -1, the whole dataset is used", + "Aliases": [ + "numauc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.AnomalyPipelineEnsemble", + "Desc": "Combine anomaly detection models into an ensemble", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Average" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IAnomalyDetectionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.BinaryClassificationEvaluator", + "Desc": "Evaluates a binary classification scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "ProbabilityColumn", + "Type": "String", + "Desc": "Probability column name", + "Aliases": [ + "prob" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Threshold", + "Type": "Float", + "Desc": "Probability value for classification thresholding", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "UseRawScoreThreshold", + "Type": "Bool", + "Desc": "Use raw score value instead of probability for classification thresholding", + "Aliases": [ + "useRawScore" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumRocExamples", + "Type": "Int", + "Desc": "The number of samples to use for p/r curve generation. Specify 0 for no p/r curve generation", + "Aliases": [ + "numpr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100000 + }, + { + "Name": "MaxAucExamples", + "Type": "Int", + "Desc": "The number of samples to use for AUC calculation. If 0, AUC is not computed. If -1, the whole dataset is used", + "Aliases": [ + "numauc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "NumAuPrcExamples", + "Type": "Int", + "Desc": "The number of samples to use for AUPRC calculation. Specify 0 for no AUPRC calculation", + "Aliases": [ + "numauprc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100000 + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + }, + { + "Name": "ConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IClassificationEvaluatorOutput", + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.BinaryEnsemble", + "Desc": "Combine binary classifiers into an ensemble", + "FriendlyName": "Parallel Ensemble (bagging, stacking, etc)", + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average", + "Vote" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Median" + }, + { + "Name": "ValidatePipelines", + "Type": "Bool", + "Desc": "Whether to validate that all the pipelines are identical", + "Aliases": [ + "validate" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.BinaryPipelineEnsemble", + "Desc": "Combine binary classification models into an ensemble", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average", + "Vote" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Median" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.ClassificationEvaluator", + "Desc": "Evaluates a multi class classification scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "OutputTopKAcc", + "Type": "Int", + "Desc": "Output top-K accuracy.", + "Aliases": [ + "topkacc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumTopClassesToOutput", + "Type": "Int", + "Desc": "Output top-K classes.", + "Aliases": [ + "topk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 3 + }, + { + "Name": "NumClassesConfusionMatrix", + "Type": "Int", + "Desc": "Maximum number of classes in confusion matrix.", + "Aliases": [ + "nccf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "OutputPerClassStatistics", + "Type": "Bool", + "Desc": "Output per class statistics and confusion matrix.", + "Aliases": [ + "opcs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + }, + { + "Name": "ConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IClassificationEvaluatorOutput", + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.ClusterEvaluator", + "Desc": "Evaluates a clustering scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "FeatureColumn", + "Type": "String", + "Desc": "Features column name", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "CalculateDbi", + "Type": "Bool", + "Desc": "Calculate DBI? (time-consuming unsupervised metric)", + "Aliases": [ + "dbi" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NumTopClustersToOutput", + "Type": "Int", + "Desc": "Output top K clusters", + "Aliases": [ + "topk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 3 + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.CrossValidationResultsCombiner", + "Desc": "Combine the metric data views returned from cross validation.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "OverallMetrics", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Overall metrics datasets", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PerInstanceMetrics", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Per instance metrics datasets", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ConfusionMatrix", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Confusion matrix datasets", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Warnings", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Warning datasets", + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Kind", + "Type": { + "Kind": "Enum", + "Values": [ + "SignatureBinaryClassifierTrainer", + "SignatureMulticlassClassificationTrainer", + "SignatureRankerTrainer", + "SignatureRegressorTrainer", + "SignatureMultiOutputRegressorTrainer", + "SignatureAnomalyDetectorTrainer", + "SignatureClusteringTrainer" + ] + }, + "Desc": "Specifies the trainer kind, which determines the evaluator to be used.", + "Required": true, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "SignatureBinaryClassifierTrainer" + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "The label column name", + "Aliases": [ + "Label" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 7.0, + "IsNullable": false, + "Default": "Weight" + }, + { + "Name": "GroupColumn", + "Type": "String", + "Desc": "Column to use for grouping", + "Aliases": [ + "group" + ], + "Required": false, + "SortOrder": 8.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 9.0, + "IsNullable": false, + "Default": "Name" + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + }, + { + "Name": "ConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset" + } + ] + }, + { + "Name": "Models.CrossValidator", + "Desc": "Cross validation for general learning", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data set", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "TransformModel", + "Type": "TransformModel", + "Desc": "The transform model from the pipeline before this command. It gets included in the Output.PredictorModel.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Nodes", + "Type": { + "Kind": "Array", + "ItemType": "Node" + }, + "Desc": "The training subgraph", + "Required": true, + "SortOrder": 3.0, + "IsNullable": false + }, + { + "Name": "Inputs", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for training", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ] + }, + "Desc": "The training subgraph inputs", + "Required": true, + "SortOrder": 4.0, + "IsNullable": false + }, + { + "Name": "Outputs", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + } + ] + }, + "Desc": "The training subgraph outputs", + "Required": true, + "SortOrder": 5.0, + "IsNullable": false + }, + { + "Name": "StratificationColumn", + "Type": "String", + "Desc": "Column to use for stratification", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NumFolds", + "Type": "Int", + "Desc": "Number of folds in k-fold cross-validation", + "Aliases": [ + "k" + ], + "Required": false, + "SortOrder": 7.0, + "IsNullable": false, + "Default": 2 + }, + { + "Name": "Kind", + "Type": { + "Kind": "Enum", + "Values": [ + "SignatureBinaryClassifierTrainer", + "SignatureMulticlassClassificationTrainer", + "SignatureRankerTrainer", + "SignatureRegressorTrainer", + "SignatureMultiOutputRegressorTrainer", + "SignatureAnomalyDetectorTrainer", + "SignatureClusteringTrainer" + ] + }, + "Desc": "Specifies the trainer kind, which determines the evaluator to be used.", + "Required": true, + "SortOrder": 8.0, + "IsNullable": false, + "Default": "SignatureBinaryClassifierTrainer" + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 9.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 10.0, + "IsNullable": false, + "Default": "Weight" + }, + { + "Name": "GroupColumn", + "Type": "String", + "Desc": "Column to use for grouping", + "Aliases": [ + "group" + ], + "Required": false, + "SortOrder": 11.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 12.0, + "IsNullable": false, + "Default": "Name" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The final model including the trained predictor model and the model from the transforms, provided as the Input.TransformModel." + }, + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + }, + { + "Name": "ConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset" + } + ] + }, + { + "Name": "Models.CrossValidatorDatasetSplitter", + "Desc": "Split the dataset into the specified number of cross-validation folds (train and test sets)", + "FriendlyName": "Dataset CV Split", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumFolds", + "Type": "Int", + "Desc": "Number of folds to split into", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 2 + }, + { + "Name": "StratificationColumn", + "Type": "String", + "Desc": "Stratification column", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "TrainData", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Training data (one dataset per fold)" + }, + { + "Name": "TestData", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Testing data (one dataset per fold)" + } + ] + }, + { + "Name": "Models.DatasetTransformer", + "Desc": "Applies a TransformModel to a dataset.", + "FriendlyName": "Apply Transform Model Output", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "TransformModel", + "Type": "TransformModel", + "Desc": "Transform model", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + } + ], + "InputKind": [ + "ITransformInput" + ] + }, + { + "Name": "Models.EnsembleSummary", + "Desc": "Summarize a pipeline ensemble predictor.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to summarize", + "Aliases": [ + "predictorModel" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "Summaries", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "The summaries of the individual predictors" + }, + { + "Name": "Stats", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "The model statistics of the individual predictors" + } + ] + }, + { + "Name": "Models.FixedPlattCalibrator", + "Desc": "Apply a Platt calibrator with a fixed slope and offset to an input model", + "FriendlyName": "Fixed Sigmoid Calibration", + "ShortName": null, + "Inputs": [ + { + "Name": "Slope", + "Type": "Float", + "Desc": "The slope parameter of the calibration function 1 / (1 + exp(slope * x + offset)", + "Aliases": [ + "slope" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": -1.0 + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UncalibratedPredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to calibrate", + "Aliases": [ + "uncalibratedPredictorModel" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "The offset parameter of the calibration function 1 / (1 + exp(slope * x + offset)", + "Aliases": [ + "offset" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "MaxRows", + "Type": "Int", + "Desc": "The maximum number of examples to train the calibrator on", + "Aliases": [ + "maxRows" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1000000000, + "Range": { + "Inf": 0, + "Max": 2147483647 + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ICalibratorInput", + "ITransformInput" + ], + "OutputKind": [ + "ICalibratorOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.MultiClassPipelineEnsemble", + "Desc": "Combine multiclass classifiers into an ensemble", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average", + "Vote" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Median" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.MultiOutputRegressionEvaluator", + "Desc": "Evaluates a multi output regression scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "RegressionLossFunction" + }, + "Desc": "Loss function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "SquaredLoss" + } + }, + { + "Name": "SuppressScoresAndLabels", + "Type": "Bool", + "Desc": "Suppress labels and scores in per-instance outputs?", + "Aliases": [ + "noScores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.NaiveCalibrator", + "Desc": "Apply a Naive calibrator to an input model", + "FriendlyName": "Naive Calibrator", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UncalibratedPredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to calibrate", + "Aliases": [ + "uncalibratedPredictorModel" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "MaxRows", + "Type": "Int", + "Desc": "The maximum number of examples to train the calibrator on", + "Aliases": [ + "maxRows" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1000000000, + "Range": { + "Inf": 0, + "Max": 2147483647 + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ICalibratorInput", + "ITransformInput" + ], + "OutputKind": [ + "ICalibratorOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.OneVersusAll", + "Desc": "One-vs-All macro (OVA)", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Nodes", + "Type": { + "Kind": "Array", + "ItemType": "Node" + }, + "Desc": "The subgraph for the binary trainer used to construct the OVA learner. This should be a TrainBinary node.", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "OutputForSubGraph", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Model", + "Type": "PredictorModel", + "Desc": "The predictor model for the subgraph exemplar.", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ] + }, + "Desc": "The training subgraph output.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "UseProbabilities", + "Type": "Bool", + "Desc": "Use probabilities in OVA combiner", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained multiclass model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ] + }, + { + "Name": "Models.OnnxConverter", + "Desc": "Converts the model to ONNX format.", + "FriendlyName": "ONNX Converter.", + "ShortName": null, + "Inputs": [ + { + "Name": "DataFile", + "Type": "String", + "Desc": "The data file", + "Aliases": [ + "data" + ], + "Required": false, + "SortOrder": 0.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Onnx", + "Type": "String", + "Desc": "The path to write the output ONNX to.", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Json", + "Type": "String", + "Desc": "The path to write the output JSON to.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The 'name' property in the output ONNX. By default this will be the ONNX extension-less name.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Domain", + "Type": "String", + "Desc": "The 'domain' property in the output ONNX.", + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "InputsToDrop", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Array of input column names to drop", + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "OutputsToDrop", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Array of output column names to drop", + "Required": false, + "SortOrder": 8.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Model that needs to be converted to ONNX format.", + "Required": false, + "SortOrder": 10.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "OnnxVersion", + "Type": { + "Kind": "Enum", + "Values": [ + "Stable", + "Experimental" + ] + }, + "Desc": "The targeted ONNX version. It can be either \"Stable\" or \"Experimental\". If \"Experimental\" is used, produced model can contain components that is not officially supported in ONNX standard.", + "Required": false, + "SortOrder": 11.0, + "IsNullable": false, + "Default": "Stable" + }, + { + "Name": "PredictiveModel", + "Type": "PredictorModel", + "Desc": "Predictor model that needs to be converted to ONNX format.", + "Required": false, + "SortOrder": 12.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [] + }, + { + "Name": "Models.OvaModelCombiner", + "Desc": "Combines a sequence of PredictorModels into a single model", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "ModelArray", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "Input models", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseProbabilities", + "Type": "Bool", + "Desc": "Use probabilities from learners instead of raw values.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Predictor model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ] + }, + { + "Name": "Models.PAVCalibrator", + "Desc": "Apply a PAV calibrator to an input model", + "FriendlyName": "PAV Calibration", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UncalibratedPredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to calibrate", + "Aliases": [ + "uncalibratedPredictorModel" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "MaxRows", + "Type": "Int", + "Desc": "The maximum number of examples to train the calibrator on", + "Aliases": [ + "maxRows" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1000000000, + "Range": { + "Inf": 0, + "Max": 2147483647 + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ICalibratorInput", + "ITransformInput" + ], + "OutputKind": [ + "ICalibratorOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.PlattCalibrator", + "Desc": "Apply a Platt calibrator to an input model", + "FriendlyName": "Sigmoid Calibration", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UncalibratedPredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to calibrate", + "Aliases": [ + "uncalibratedPredictorModel" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "MaxRows", + "Type": "Int", + "Desc": "The maximum number of examples to train the calibrator on", + "Aliases": [ + "maxRows" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1000000000, + "Range": { + "Inf": 0, + "Max": 2147483647 + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ICalibratorInput", + "ITransformInput" + ], + "OutputKind": [ + "ICalibratorOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.QuantileRegressionEvaluator", + "Desc": "Evaluates a quantile regression scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "RegressionLossFunction" + }, + "Desc": "Loss function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "SquaredLoss" + } + }, + { + "Name": "Index", + "Type": "Int", + "Desc": "Quantile index to select", + "Aliases": [ + "ind" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.RankingEvaluator", + "Desc": "Evaluates a ranking scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "GroupIdColumn", + "Type": "String", + "Desc": "Column to use for the group ID", + "Aliases": [ + "group" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "DcgTruncationLevel", + "Type": "Int", + "Desc": "Maximum truncation level for computing (N)DCG", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "LabelGains", + "Type": "String", + "Desc": "Label relevance gains", + "Aliases": [ + "gains" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "0,3,7,15,31" + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.RegressionEnsemble", + "Desc": "Combine regression models into an ensemble", + "FriendlyName": "Regression Ensemble (bagging, stacking, etc)", + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Median" + }, + { + "Name": "ValidatePipelines", + "Type": "Bool", + "Desc": "Whether to validate that all the pipelines are identical", + "Aliases": [ + "validate" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.RegressionEvaluator", + "Desc": "Evaluates a regression scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "RegressionLossFunction" + }, + "Desc": "Loss function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "SquaredLoss" + } + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.RegressionPipelineEnsemble", + "Desc": "Combine regression models into an ensemble", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Median" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.Summarizer", + "Desc": "Summarize a linear regression predictor.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to summarize", + "Aliases": [ + "predictorModel" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "Summary", + "Type": "DataView", + "Desc": "The summary of a predictor" + }, + { + "Name": "Stats", + "Type": "DataView", + "Desc": "The training set statistics. Note that this output can be null." + } + ] + }, + { + "Name": "Models.TrainTestEvaluator", + "Desc": "General train test for any supported evaluator", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "train" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "TestingData", + "Type": "DataView", + "Desc": "The data to be used for testing", + "Aliases": [ + "test" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "TransformModel", + "Type": "TransformModel", + "Desc": "The aggregated transform model from the pipeline before this command, to apply to the test data, and also include in the final model, together with the predictor model.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Nodes", + "Type": { + "Kind": "Array", + "ItemType": "Node" + }, + "Desc": "The training subgraph", + "Required": true, + "SortOrder": 4.0, + "IsNullable": false + }, + { + "Name": "Inputs", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for training", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ] + }, + "Desc": "The training subgraph inputs", + "Required": true, + "SortOrder": 5.0, + "IsNullable": false + }, + { + "Name": "Outputs", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + } + ] + }, + "Desc": "The training subgraph outputs", + "Required": true, + "SortOrder": 6.0, + "IsNullable": false + }, + { + "Name": "Kind", + "Type": { + "Kind": "Enum", + "Values": [ + "SignatureBinaryClassifierTrainer", + "SignatureMulticlassClassificationTrainer", + "SignatureRankerTrainer", + "SignatureRegressorTrainer", + "SignatureMultiOutputRegressorTrainer", + "SignatureAnomalyDetectorTrainer", + "SignatureClusteringTrainer" + ] + }, + "Desc": "Specifies the trainer kind, which determines the evaluator to be used.", + "Required": false, + "SortOrder": 7.0, + "IsNullable": false, + "Default": "SignatureBinaryClassifierTrainer" + }, + { + "Name": "PipelineId", + "Type": "String", + "Desc": "Identifies which pipeline was run for this train test.", + "Required": false, + "SortOrder": 8.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "IncludeTrainingMetrics", + "Type": "Bool", + "Desc": "Indicates whether to include and output training dataset metrics.", + "Required": false, + "SortOrder": 9.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 10.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 11.0, + "IsNullable": false, + "Default": "Weight" + }, + { + "Name": "GroupColumn", + "Type": "String", + "Desc": "Column to use for grouping", + "Aliases": [ + "group" + ], + "Required": false, + "SortOrder": 12.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 13.0, + "IsNullable": false, + "Default": "Name" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The final model including the trained predictor model and the model from the transforms, provided as the Input.TransformModel." + }, + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + }, + { + "Name": "ConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset" + }, + { + "Name": "TrainingWarnings", + "Type": "DataView", + "Desc": "Warning dataset for training" + }, + { + "Name": "TrainingOverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset for training" + }, + { + "Name": "TrainingPerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset for training" + }, + { + "Name": "TrainingConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset for training" + } + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.ExponentialAverage", + "Desc": "Applies a Exponential average on a time series.", + "FriendlyName": "Exponential Average Transform", + "ShortName": "ExpAvg", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column", + "Aliases": [ + "name" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Decay", + "Type": "Float", + "Desc": "Coefficient d in: d m(y_t) = d * y_t + (1-d) * m(y_(t-1)), it should be in [0, 1].", + "Aliases": [ + "d" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.9 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.IidChangePointDetector", + "Desc": "This transform detects the change-points in an i.i.d. sequence using adaptive kernel density estimation and martingales.", + "FriendlyName": "IID Change Point Detection", + "ShortName": "ichgpnt", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column.", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Confidence", + "Type": "Float", + "Desc": "The confidence for change point detection in the range [0, 100].", + "Aliases": [ + "cnf" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 95.0 + }, + { + "Name": "ChangeHistoryLength", + "Type": "Int", + "Desc": "The length of the sliding window on p-values for computing the martingale score.", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": 20 + }, + { + "Name": "Martingale", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Power", + "Mixture" + ] + }, + "Desc": "The martingale used for scoring.", + "Aliases": [ + "mart" + ], + "Required": false, + "SortOrder": 103.0, + "IsNullable": false, + "Default": "Power" + }, + { + "Name": "PowerMartingaleEpsilon", + "Type": "Float", + "Desc": "The epsilon parameter for the Power martingale.", + "Aliases": [ + "eps" + ], + "Required": false, + "SortOrder": 104.0, + "IsNullable": false, + "Default": 0.1 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.IidSpikeDetector", + "Desc": "This transform detects the spikes in a i.i.d. sequence using adaptive kernel density estimation.", + "FriendlyName": "IID Spike Detection", + "ShortName": "ispike", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column.", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Confidence", + "Type": "Float", + "Desc": "The confidence for spike detection in the range [0, 100].", + "Aliases": [ + "cnf" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 99.0 + }, + { + "Name": "Side", + "Type": { + "Kind": "Enum", + "Values": [ + "Positive", + "Negative", + "TwoSided" + ] + }, + "Desc": "The argument that determines whether to detect positive or negative anomalies, or both.", + "Aliases": [ + "side" + ], + "Required": false, + "SortOrder": 101.0, + "IsNullable": false, + "Default": "TwoSided" + }, + { + "Name": "PvalueHistoryLength", + "Type": "Int", + "Desc": "The size of the sliding window for computing the p-value.", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": 100 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.PercentileThresholdTransform", + "Desc": "Detects the values of time-series that are in the top percentile of the sliding window.", + "FriendlyName": "Percentile Threshold Transform", + "ShortName": "TopPcnt", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column", + "Aliases": [ + "name" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Percentile", + "Type": "Float", + "Desc": "The percentile value for thresholding in the range [0, 100]", + "Aliases": [ + "pcnt" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The size of the sliding window for computing the percentile threshold. The default value is set to 1.", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 1 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.PValueTransform", + "Desc": "This P-Value transform calculates the p-value of the current input in the sequence with regard to the values in the sliding window.", + "FriendlyName": "p-Value Transform", + "ShortName": "PVal", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column", + "Aliases": [ + "name" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed value of the random generator", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "PositiveSide", + "Type": "Bool", + "Desc": "The flag that determines whether the p-values are calculated on the positive side", + "Aliases": [ + "pos" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The size of the sliding window for computing the p-value", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "InitialWindowSize", + "Type": "Int", + "Desc": "The size of the initial window for computing the p-value. The default value is set to 0, which means there is no initial window considered.", + "Aliases": [ + "initwnd" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.SlidingWindowTransform", + "Desc": "Returns the last values for a time series [y(t-d-l+1), y(t-d-l+2), ..., y(t-l-1), y(t-l)] where d is the size of the window, l the lag and y is a Float.", + "FriendlyName": "Sliding Window Transform", + "ShortName": "SlideWin", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The size of the sliding window for computing the moving average", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 2 + }, + { + "Name": "Lag", + "Type": "Int", + "Desc": "Lag between current observation and last observation from the sliding window", + "Aliases": [ + "l" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "Begin", + "Type": { + "Kind": "Enum", + "Values": [ + "NaNValues", + "FirstValue" + ] + }, + "Desc": "Define how to populate the first rows of the produced series", + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "NaNValues" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.SsaChangePointDetector", + "Desc": "This transform detects the change-points in a seasonal time-series using Singular Spectrum Analysis (SSA).", + "FriendlyName": "SSA Change Point Detection", + "ShortName": "chgpnt", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column.", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "TrainingWindowSize", + "Type": "Int", + "Desc": "The number of points from the beginning of the sequence used for training.", + "Aliases": [ + "twnd" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Confidence", + "Type": "Float", + "Desc": "The confidence for change point detection in the range [0, 100].", + "Aliases": [ + "cnf" + ], + "Required": true, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 95.0 + }, + { + "Name": "SeasonalWindowSize", + "Type": "Int", + "Desc": "An upper bound on the largest relevant seasonality in the input time-series.", + "Aliases": [ + "swnd" + ], + "Required": true, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "ChangeHistoryLength", + "Type": "Int", + "Desc": "The length of the sliding window on p-values for computing the martingale score.", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": 20 + }, + { + "Name": "ErrorFunction", + "Type": { + "Kind": "Enum", + "Values": [ + "SignedDifference", + "AbsoluteDifference", + "SignedProportion", + "AbsoluteProportion", + "SquaredDifference" + ] + }, + "Desc": "The function used to compute the error between the expected and the observed value.", + "Aliases": [ + "err" + ], + "Required": false, + "SortOrder": 103.0, + "IsNullable": false, + "Default": "SignedDifference" + }, + { + "Name": "Martingale", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Power", + "Mixture" + ] + }, + "Desc": "The martingale used for scoring.", + "Aliases": [ + "mart" + ], + "Required": false, + "SortOrder": 104.0, + "IsNullable": false, + "Default": "Power" + }, + { + "Name": "PowerMartingaleEpsilon", + "Type": "Float", + "Desc": "The epsilon parameter for the Power martingale.", + "Aliases": [ + "eps" + ], + "Required": false, + "SortOrder": 105.0, + "IsNullable": false, + "Default": 0.1 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.SsaForecasting", + "Desc": "This transform forecasts using Singular Spectrum Analysis (SSA).", + "FriendlyName": "SSA Forecasting", + "ShortName": "ssafcst", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column.", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The length of the window on the series for building the trajectory matrix (parameter L).", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "SeriesLength", + "Type": "Int", + "Desc": "The length of series that is kept in buffer for modeling (parameter N).", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "TrainSize", + "Type": "Int", + "Desc": "The length of series from the beginning used for training.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Horizon", + "Type": "Int", + "Desc": "The number of values to forecast.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "ConfidenceLevel", + "Type": "Float", + "Desc": "The confidence level in [0, 1) for forecasting.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0.95 + }, + { + "Name": "VariableHorizon", + "Type": "Bool", + "Desc": "Set this to true horizon will change at prediction time.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "ConfidenceLowerBoundColumn", + "Type": "String", + "Desc": "The name of the confidence interval lower bound column.", + "Aliases": [ + "cnfminname" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ConfidenceUpperBoundColumn", + "Type": "String", + "Desc": "The name of the confidence interval upper bound column.", + "Aliases": [ + "cnfmaxnname" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RankSelectionMethod", + "Type": { + "Kind": "Enum", + "Values": [ + "Fixed", + "Exact", + "Fast" + ] + }, + "Desc": "The rank selection method.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Exact" + }, + { + "Name": "Rank", + "Type": "Int", + "Desc": "The desired rank of the subspace used for SSA projection (parameter r). This parameter should be in the range in [1, windowSize]. If set to null, the rank is automatically determined based on prediction error minimization.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxRank", + "Type": "Int", + "Desc": "The maximum rank considered during the rank selection process. If not provided (i.e. set to null), it is set to windowSize - 1.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ShouldStabilize", + "Type": "Bool", + "Desc": "The flag determining whether the model should be stabilized.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "ShouldMaintainInfo", + "Type": "Bool", + "Desc": "The flag determining whether the meta information for the model needs to be maintained.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaxGrowth", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "TimeSpan", + "Type": "Int", + "Desc": "Time span of growth ratio. Must be strictly positive.", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Growth", + "Type": "Float", + "Desc": "Growth. Must be non-negative.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0.0 + } + ] + }, + "Desc": "The maximum growth on the exponential trend.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DiscountFactor", + "Type": "Float", + "Desc": "The discount factor in [0,1] used for online updates.", + "Aliases": [ + "disc" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "IsAdaptive", + "Type": "Bool", + "Desc": "The flag determing whether the model is adaptive", + "Aliases": [ + "adp" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.SsaSpikeDetector", + "Desc": "This transform detects the spikes in a seasonal time-series using Singular Spectrum Analysis (SSA).", + "FriendlyName": "SSA Spike Detection", + "ShortName": "spike", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column.", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "TrainingWindowSize", + "Type": "Int", + "Desc": "The number of points from the beginning of the sequence used for training.", + "Aliases": [ + "twnd" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Confidence", + "Type": "Float", + "Desc": "The confidence for spike detection in the range [0, 100].", + "Aliases": [ + "cnf" + ], + "Required": true, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 99.0 + }, + { + "Name": "SeasonalWindowSize", + "Type": "Int", + "Desc": "An upper bound on the largest relevant seasonality in the input time-series.", + "Aliases": [ + "swnd" + ], + "Required": true, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "Side", + "Type": { + "Kind": "Enum", + "Values": [ + "Positive", + "Negative", + "TwoSided" + ] + }, + "Desc": "The argument that determines whether to detect positive or negative anomalies, or both.", + "Aliases": [ + "side" + ], + "Required": false, + "SortOrder": 101.0, + "IsNullable": false, + "Default": "TwoSided" + }, + { + "Name": "PvalueHistoryLength", + "Type": "Int", + "Desc": "The size of the sliding window for computing the p-value.", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "ErrorFunction", + "Type": { + "Kind": "Enum", + "Values": [ + "SignedDifference", + "AbsoluteDifference", + "SignedProportion", + "AbsoluteProportion", + "SquaredDifference" + ] + }, + "Desc": "The function used to compute the error between the expected and the observed value.", + "Aliases": [ + "err" + ], + "Required": false, + "SortOrder": 103.0, + "IsNullable": false, + "Default": "SignedDifference" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Trainers.AveragedPerceptronBinaryClassifier", + "Desc": "Averaged Perceptron Binary Classifier.", + "FriendlyName": "Averaged Perceptron", + "ShortName": "ap", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "ClassificationLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "HingeLoss" + } + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.01, + 0.1, + 0.5, + 1.0 + ] + } + }, + { + "Name": "DecreaseLearningRate", + "Type": "Bool", + "Desc": "Decrease learning rate", + "Aliases": [ + "decreaselr" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 Regularization Weight", + "Aliases": [ + "reg", + "L2RegularizerWeight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 0.4 + } + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter", + "numIterations" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 100, + "StepSize": 10.0, + "IsLogScale": true + } + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "initWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "ResetWeightsAfterXExamples", + "Type": "Int", + "Desc": "Number of examples after which weights will be reset to the current average", + "Aliases": [ + "numreset" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LazyUpdate", + "Type": "Bool", + "Desc": "Instead of updating averaged weights on every example, only update when loss is nonzero", + "Aliases": [ + "lazy", + "DoLazyUpdates" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "RecencyGain", + "Type": "Float", + "Desc": "Extra weight given to more recent updates", + "Aliases": [ + "rg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "RecencyGainMultiplicative", + "Type": "Bool", + "Desc": "Whether Recency Gain is multiplicative (vs. additive)", + "Aliases": [ + "rgm", + "RecencyGainMulti" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Averaged", + "Type": "Bool", + "Desc": "Do averaging?", + "Aliases": [ + "avg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "AveragedTolerance", + "Type": "Float", + "Desc": "The inexactness tolerance for averaging", + "Aliases": [ + "avgtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01 + }, + { + "Name": "InitialWeights", + "Type": "String", + "Desc": "Initial Weights and bias, comma-separated", + "Aliases": [ + "initweights" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Whether to shuffle for each training iteration", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.EnsembleBinaryClassifier", + "Desc": "Train binary ensemble.", + "FriendlyName": "Parallel Ensemble (bagging, stacking, etc)", + "ShortName": null, + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "SamplingType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleSubsetSelector" + }, + "Desc": "Sampling Type", + "Aliases": [ + "st" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": { + "Name": "BootstrapSelector", + "Settings": { + "FeatureSelector": { + "Name": "AllFeatureSelector" + } + } + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "NumModels", + "Type": "Int", + "Desc": "Number of models per batch. If not specified, will default to 50 if there is only one base predictor, or the number of base predictors otherwise.", + "Aliases": [ + "nm" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "SubModelSelectorType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleBinarySubModelSelector" + }, + "Desc": "Algorithm to prune the base learners for selective Ensemble", + "Aliases": [ + "pt" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": { + "Name": "AllSelector" + } + }, + { + "Name": "OutputCombiner", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleBinaryOutputCombiner" + }, + "Desc": "Output combiner", + "Aliases": [ + "oc" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": { + "Name": "Median" + } + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "TrainParallel", + "Type": "Bool", + "Desc": "All the base learners will run asynchronously if the value is true", + "Aliases": [ + "tp" + ], + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Batch size", + "Aliases": [ + "bs" + ], + "Required": false, + "SortOrder": 107.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "ShowMetrics", + "Type": "Bool", + "Desc": "True, if metrics for each model need to be evaluated and shown in comparison table. This is done by using validation set if available or the training set", + "Aliases": [ + "sm" + ], + "Required": false, + "SortOrder": 108.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.EnsembleClassification", + "Desc": "Train multiclass ensemble.", + "FriendlyName": "Parallel Ensemble (bagging, stacking, etc)", + "ShortName": null, + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "SamplingType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleSubsetSelector" + }, + "Desc": "Sampling Type", + "Aliases": [ + "st" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": { + "Name": "BootstrapSelector", + "Settings": { + "FeatureSelector": { + "Name": "AllFeatureSelector" + } + } + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "NumModels", + "Type": "Int", + "Desc": "Number of models per batch. If not specified, will default to 50 if there is only one base predictor, or the number of base predictors otherwise.", + "Aliases": [ + "nm" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "SubModelSelectorType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleMulticlassSubModelSelector" + }, + "Desc": "Algorithm to prune the base learners for selective Ensemble", + "Aliases": [ + "pt" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": { + "Name": "AllSelectorMultiClass" + } + }, + { + "Name": "OutputCombiner", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleMulticlassOutputCombiner" + }, + "Desc": "Output combiner", + "Aliases": [ + "oc" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": { + "Name": "MultiMedian" + } + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "TrainParallel", + "Type": "Bool", + "Desc": "All the base learners will run asynchronously if the value is true", + "Aliases": [ + "tp" + ], + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Batch size", + "Aliases": [ + "bs" + ], + "Required": false, + "SortOrder": 107.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "ShowMetrics", + "Type": "Bool", + "Desc": "True, if metrics for each model need to be evaluated and shown in comparison table. This is done by using validation set if available or the training set", + "Aliases": [ + "sm" + ], + "Required": false, + "SortOrder": 108.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.EnsembleRegression", + "Desc": "Train regression ensemble.", + "FriendlyName": "Parallel Ensemble (bagging, stacking, etc)", + "ShortName": null, + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "SamplingType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleSubsetSelector" + }, + "Desc": "Sampling Type", + "Aliases": [ + "st" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": { + "Name": "BootstrapSelector", + "Settings": { + "FeatureSelector": { + "Name": "AllFeatureSelector" + } + } + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "NumModels", + "Type": "Int", + "Desc": "Number of models per batch. If not specified, will default to 50 if there is only one base predictor, or the number of base predictors otherwise.", + "Aliases": [ + "nm" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "SubModelSelectorType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleRegressionSubModelSelector" + }, + "Desc": "Algorithm to prune the base learners for selective Ensemble", + "Aliases": [ + "pt" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": { + "Name": "AllSelector" + } + }, + { + "Name": "OutputCombiner", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleRegressionOutputCombiner" + }, + "Desc": "Output combiner", + "Aliases": [ + "oc" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": { + "Name": "Median" + } + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "TrainParallel", + "Type": "Bool", + "Desc": "All the base learners will run asynchronously if the value is true", + "Aliases": [ + "tp" + ], + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Batch size", + "Aliases": [ + "bs" + ], + "Required": false, + "SortOrder": 107.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "ShowMetrics", + "Type": "Bool", + "Desc": "True, if metrics for each model need to be evaluated and shown in comparison table. This is done by using validation set if available or the training set", + "Aliases": [ + "sm" + ], + "Required": false, + "SortOrder": 108.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastForestBinaryClassifier", + "Desc": "Uses a random forest learner to perform binary classification.", + "FriendlyName": "Fast Forest Classification", + "ShortName": "ff", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "MaximumOutputMagnitudePerTree", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "NumberOfQuantileSamples", + "Type": "Int", + "Desc": "Number of labels to be sampled from each leaf to make the distribution", + "Aliases": [ + "qsc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastForestRegressor", + "Desc": "Trains a random forest to fit target values using least-squares.", + "FriendlyName": "FastForestRegression", + "ShortName": "ffr", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "ShuffleLabels", + "Type": "Bool", + "Desc": "Shuffle the labels on every iteration. Useful probably only if using this tree as a tree leaf featurizer for multiclass.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NumberOfQuantileSamples", + "Type": "Int", + "Desc": "Number of labels to be sampled from each leaf to make the distribution", + "Aliases": [ + "qsc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastTreeBinaryClassifier", + "Desc": "Uses a logit-boost boosted tree learner to perform binary classification.", + "FriendlyName": "FastTree (Boosted Trees) Classification", + "ShortName": "ftc", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "UnbalancedSets", + "Type": "Bool", + "Desc": "Option for using derivatives optimized for unbalanced sets", + "Aliases": [ + "us" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastTreeRanker", + "Desc": "Trains gradient boosted decision trees to the LambdaRank quasi-gradient.", + "FriendlyName": "FastTree (Boosted Trees) Ranking", + "ShortName": "ftrank", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "CustomGains", + "Type": { + "Kind": "Array", + "ItemType": "Float" + }, + "Desc": "Comma-separated list of gains associated to each relevance label.", + "Aliases": [ + "gains" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + 0.0, + 3.0, + 7.0, + 15.0, + 31.0 + ] + }, + { + "Name": "UseDcg", + "Type": "Bool", + "Desc": "Train DCG instead of NDCG", + "Aliases": [ + "dcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "SortingAlgorithm", + "Type": "String", + "Desc": "The sorting algorithm to use for DCG and LambdaMart calculations [DescendingStablePessimistic/DescendingStable/DescendingReverse/DescendingDotNet]", + "Aliases": [ + "sort" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "DescendingStablePessimistic" + }, + { + "Name": "NdcgTruncationLevel", + "Type": "Int", + "Desc": "max-NDCG truncation to use in the LambdaMART algorithm", + "Aliases": [ + "n" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "ShiftedNdcg", + "Type": "Bool", + "Desc": "Use shifted NDCG", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "CostFunctionParam", + "Type": "Char", + "Desc": "Cost function parameter (w/c)", + "Aliases": [ + "cf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "w" + }, + { + "Name": "DistanceWeight2", + "Type": "Bool", + "Desc": "Distance weight 2 adjustment to cost", + "Aliases": [ + "dw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NormalizeQueryLambdas", + "Type": "Bool", + "Desc": "Normalize query lambdas", + "Aliases": [ + "nql" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRankingOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastTreeRegressor", + "Desc": "Trains gradient boosted decision trees to fit target values using least-squares.", + "FriendlyName": "FastTree (Boosted Trees) Regression", + "ShortName": "ftr", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastTreeTweedieRegressor", + "Desc": "Trains gradient boosted decision trees to fit target values using a Tweedie loss function. This learner is a generalization of Poisson, compound Poisson, and gamma regression.", + "FriendlyName": "FastTree (Boosted Trees) Tweedie Regression", + "ShortName": "fttweedie", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Index", + "Type": "Float", + "Desc": "Index parameter for the Tweedie distribution, in the range [1, 2]. 1 is Poisson loss, 2 is gamma loss, and intermediate values are compound Poisson loss.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.5 + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FieldAwareFactorizationMachineBinaryClassifier", + "Desc": "Train a field-aware factorization machine for binary classification", + "FriendlyName": "Field-aware Factorization Machine", + "ShortName": "ffm", + "Inputs": [ + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Initial learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 0.1, + "SweepRange": { + "RangeType": "Float", + "Min": 0.001, + "Max": 1.0, + "IsLogScale": true + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of training iterations", + "Aliases": [ + "iters", + "iter" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 5, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 100 + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LatentDimension", + "Type": "Int", + "Desc": "Latent space dimension", + "Aliases": [ + "d" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 4, + "Max": 100 + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LambdaLinear", + "Type": "Float", + "Desc": "Regularization coefficient of linear weights", + "Aliases": [ + "lambdaLinear" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.0001, + "SweepRange": { + "RangeType": "Float", + "Min": 1E-08, + "Max": 1.0, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "LambdaLatent", + "Type": "Float", + "Desc": "Regularization coefficient of latent weights", + "Aliases": [ + "lambdaLatent" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 0.0001, + "SweepRange": { + "RangeType": "Float", + "Min": 1E-08, + "Max": 1.0, + "IsLogScale": true + } + }, + { + "Name": "NormalizeFeatures", + "Type": "Bool", + "Desc": "Whether to normalize the input vectors so that the concatenation of all fields' feature vectors is unit-length", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "ExtraFeatureColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Extra columns to use for feature vectors. The i-th specified string denotes the column containing features form the (i+1)-th field. Note that the first field is specified by \"feat\" instead of \"exfeat\".", + "Aliases": [ + "exfeat" + ], + "Required": false, + "SortOrder": 7.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Whether to shuffle for each training iteration", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 90.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Verbose", + "Type": "Bool", + "Desc": "Report traning progress or not", + "Aliases": [ + "verbose" + ], + "Required": false, + "SortOrder": 91.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Radius", + "Type": "Float", + "Desc": "Radius of initial latent factors", + "Aliases": [ + "rad" + ], + "Required": false, + "SortOrder": 110.0, + "IsNullable": false, + "Default": 0.5, + "SweepRange": { + "RangeType": "Float", + "Min": 0.1, + "Max": 1.0 + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.GeneralizedAdditiveModelBinaryClassifier", + "Desc": "Trains a gradient boosted stump per feature, on all features simultaneously, to fit target values using least-squares. It maintains no interactions between features.", + "FriendlyName": "Generalized Additive Model for Binary Classification", + "ShortName": "gam", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Total number of iterations over all features", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 9500, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 200, + 1500, + 9500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of training instances required to form a partition", + "Aliases": [ + "mi" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.002, + "SweepRange": { + "RangeType": "Float", + "Min": 0.001, + "Max": 0.1, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "UnbalancedSets", + "Type": "Bool", + "Desc": "Should we use derivatives optimized for unbalanced sets", + "Aliases": [ + "us" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Int", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Infinity" + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.GeneralizedAdditiveModelRegressor", + "Desc": "Trains a gradient boosted stump per feature, on all features simultaneously, to fit target values using least-squares. It maintains no interactions between features.", + "FriendlyName": "Generalized Additive Model for Regression", + "ShortName": "gamr", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Total number of iterations over all features", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 9500, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 200, + 1500, + 9500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of training instances required to form a partition", + "Aliases": [ + "mi" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.002, + "SweepRange": { + "RangeType": "Float", + "Min": 0.001, + "Max": 0.1, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "PruningMetrics", + "Type": "Int", + "Desc": "Metric for pruning. (For regression, 1: L1, 2:L2; default L2)", + "Aliases": [ + "pmetric" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Int", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Infinity" + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.KMeansPlusPlusClusterer", + "Desc": "K-means is a popular clustering algorithm. With K-means, the data is clustered into a specified number of clusters in order to minimize the within-cluster sum of squares. K-means++ improves upon K-means by using a better method for choosing the initial cluster centers.", + "FriendlyName": "KMeans++ Clustering", + "ShortName": "KM", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "K", + "Type": "Int", + "Desc": "The number of clusters", + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 5, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 5, + 10, + 20, + 40 + ] + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Defaults to automatic. Determinism not guaranteed.", + "Aliases": [ + "nt", + "t", + "threads" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "InitializationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "KMeansPlusPlus", + "Random", + "KMeansYinyang" + ] + }, + "Desc": "Cluster initialization algorithm", + "Aliases": [ + "init" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "KMeansYinyang" + }, + { + "Name": "OptTol", + "Type": "Float", + "Desc": "Tolerance parameter for trainer convergence. Low = slower, more accurate", + "Aliases": [ + "ot" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1E-07 + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum number of iterations.", + "Aliases": [ + "maxiter", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000 + }, + { + "Name": "AccelMemBudgetMb", + "Type": "Int", + "Desc": "Memory budget (in MBs) to use for KMeans acceleration", + "Aliases": [ + "accelMemBudgetMb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 4096 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "IUnsupervisedTrainerWithWeight", + "ITrainerInput" + ], + "OutputKind": [ + "IClusteringOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LightGbmBinaryClassifier", + "Desc": "Train a LightGBM binary classification model.", + "FriendlyName": "LightGBM Binary Classifier", + "ShortName": "LightGBM", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations.", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 50, + 100, + 150, + 200 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Shrinkage rate for trees, used to prevent over-fitting. Range: (0,1].", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "Maximum leaves for trees.", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of instances needed in a child.", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20, + 50 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "Booster", + "Type": { + "Kind": "Component", + "ComponentKind": "BoosterParameterFunction" + }, + "Desc": "Which booster to use, can be gbtree, gblinear or dart. gbtree and dart use tree based model while gblinear uses linear function.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": { + "Name": "gbdt" + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "UnbalancedSets", + "Type": "Bool", + "Desc": "Use for binary classification when training data is not balanced.", + "Aliases": [ + "us" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "WeightOfPositiveExamples", + "Type": "Float", + "Desc": "Control the balance of positive and negative weights, useful for unbalanced classes. A typical value to consider: sum(negative cases) / sum(positive cases).", + "Aliases": [ + "ScalePosWeight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Sigmoid", + "Type": "Float", + "Desc": "Parameter for the sigmoid function.", + "Aliases": [ + "sigmoid" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "EvaluationMetric", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Default", + "Logloss", + "Error", + "AreaUnderCurve" + ] + }, + "Desc": "Evaluation metrics.", + "Aliases": [ + "em" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Logloss" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of bucket bin for features.", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "Verbose", + "Type": "Bool", + "Desc": "Verbose", + "Aliases": [ + "v" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Silent", + "Type": "Bool", + "Desc": "Printing running messages.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of parallel threads used to run LightGBM.", + "Aliases": [ + "nt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "EarlyStoppingRound", + "Type": "Int", + "Desc": "Rounds of early stopping, 0 will disable it.", + "Aliases": [ + "es" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of entries in a batch when loading data.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1048576 + }, + { + "Name": "UseCategoricalSplit", + "Type": "Bool", + "Desc": "Enable categorical split or not.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "HandleMissingValue", + "Type": "Bool", + "Desc": "Enable special handling of missing value or not.", + "Aliases": [ + "hmv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "UseZeroAsMissingValue", + "Type": "Bool", + "Desc": "Enable usage of zero (0) as missing value.", + "Aliases": [ + "uzam" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "MinimumExampleCountPerGroup", + "Type": "Int", + "Desc": "Minimum number of instances per categorical group.", + "Aliases": [ + "mdpg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 50, + 100, + 200 + ] + } + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Max number of categorical thresholds.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 32, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 8, + 16, + 32, + 64 + ] + } + }, + { + "Name": "CategoricalSmoothing", + "Type": "Float", + "Desc": "Lapalace smooth term in categorical feature spilt. Avoid the bias of small categories.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20 + ] + } + }, + { + "Name": "L2CategoricalRegularization", + "Type": "Float", + "Desc": "L2 Regularization for categorical split.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.5, + 1, + 5, + 10 + ] + } + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Sets the random seed for LightGBM to use.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelLightGBM" + }, + "Desc": "Parallel LightGBM Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LightGbmClassifier", + "Desc": "Train a LightGBM multi class model.", + "FriendlyName": "LightGBM Multi Class Classifier", + "ShortName": "LightGBMMC", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations.", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 50, + 100, + 150, + 200 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Shrinkage rate for trees, used to prevent over-fitting. Range: (0,1].", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "Maximum leaves for trees.", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of instances needed in a child.", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20, + 50 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "Booster", + "Type": { + "Kind": "Component", + "ComponentKind": "BoosterParameterFunction" + }, + "Desc": "Which booster to use, can be gbtree, gblinear or dart. gbtree and dart use tree based model while gblinear uses linear function.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": { + "Name": "gbdt" + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "UnbalancedSets", + "Type": "Bool", + "Desc": "Use for multi-class classification when training data is not balanced", + "Aliases": [ + "us" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseSoftmax", + "Type": "Bool", + "Desc": "Use softmax loss for the multi classification.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "Sigmoid", + "Type": "Float", + "Desc": "Parameter for the sigmoid function.", + "Aliases": [ + "sigmoid" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "EvaluationMetric", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Default", + "Error", + "LogLoss" + ] + }, + "Desc": "Evaluation metrics.", + "Aliases": [ + "em" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Error" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of bucket bin for features.", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "Verbose", + "Type": "Bool", + "Desc": "Verbose", + "Aliases": [ + "v" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Silent", + "Type": "Bool", + "Desc": "Printing running messages.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of parallel threads used to run LightGBM.", + "Aliases": [ + "nt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "EarlyStoppingRound", + "Type": "Int", + "Desc": "Rounds of early stopping, 0 will disable it.", + "Aliases": [ + "es" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of entries in a batch when loading data.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1048576 + }, + { + "Name": "UseCategoricalSplit", + "Type": "Bool", + "Desc": "Enable categorical split or not.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "HandleMissingValue", + "Type": "Bool", + "Desc": "Enable special handling of missing value or not.", + "Aliases": [ + "hmv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "UseZeroAsMissingValue", + "Type": "Bool", + "Desc": "Enable usage of zero (0) as missing value.", + "Aliases": [ + "uzam" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "MinimumExampleCountPerGroup", + "Type": "Int", + "Desc": "Minimum number of instances per categorical group.", + "Aliases": [ + "mdpg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 50, + 100, + 200 + ] + } + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Max number of categorical thresholds.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 32, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 8, + 16, + 32, + 64 + ] + } + }, + { + "Name": "CategoricalSmoothing", + "Type": "Float", + "Desc": "Lapalace smooth term in categorical feature spilt. Avoid the bias of small categories.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20 + ] + } + }, + { + "Name": "L2CategoricalRegularization", + "Type": "Float", + "Desc": "L2 Regularization for categorical split.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.5, + 1, + 5, + 10 + ] + } + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Sets the random seed for LightGBM to use.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelLightGBM" + }, + "Desc": "Parallel LightGBM Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LightGbmRanker", + "Desc": "Train a LightGBM ranking model.", + "FriendlyName": "LightGBM Ranking", + "ShortName": "LightGBMRank", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations.", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 50, + 100, + 150, + 200 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Shrinkage rate for trees, used to prevent over-fitting. Range: (0,1].", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "Maximum leaves for trees.", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of instances needed in a child.", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20, + 50 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "Booster", + "Type": { + "Kind": "Component", + "ComponentKind": "BoosterParameterFunction" + }, + "Desc": "Which booster to use, can be gbtree, gblinear or dart. gbtree and dart use tree based model while gblinear uses linear function.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": { + "Name": "gbdt" + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "CustomGains", + "Type": { + "Kind": "Array", + "ItemType": "Int" + }, + "Desc": "An array of gains associated to each relevance label.", + "Aliases": [ + "gains" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + 0, + 3, + 7, + 15, + 31, + 63, + 127, + 255, + 511, + 1023, + 2047, + 4095 + ] + }, + { + "Name": "Sigmoid", + "Type": "Float", + "Desc": "Parameter for the sigmoid function.", + "Aliases": [ + "sigmoid" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "EvaluationMetric", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Default", + "MeanAveragedPrecision", + "NormalizedDiscountedCumulativeGain" + ] + }, + "Desc": "Evaluation metrics.", + "Aliases": [ + "em" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "NormalizedDiscountedCumulativeGain" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of bucket bin for features.", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "Verbose", + "Type": "Bool", + "Desc": "Verbose", + "Aliases": [ + "v" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Silent", + "Type": "Bool", + "Desc": "Printing running messages.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of parallel threads used to run LightGBM.", + "Aliases": [ + "nt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "EarlyStoppingRound", + "Type": "Int", + "Desc": "Rounds of early stopping, 0 will disable it.", + "Aliases": [ + "es" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of entries in a batch when loading data.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1048576 + }, + { + "Name": "UseCategoricalSplit", + "Type": "Bool", + "Desc": "Enable categorical split or not.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "HandleMissingValue", + "Type": "Bool", + "Desc": "Enable special handling of missing value or not.", + "Aliases": [ + "hmv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "UseZeroAsMissingValue", + "Type": "Bool", + "Desc": "Enable usage of zero (0) as missing value.", + "Aliases": [ + "uzam" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "MinimumExampleCountPerGroup", + "Type": "Int", + "Desc": "Minimum number of instances per categorical group.", + "Aliases": [ + "mdpg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 50, + 100, + 200 + ] + } + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Max number of categorical thresholds.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 32, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 8, + 16, + 32, + 64 + ] + } + }, + { + "Name": "CategoricalSmoothing", + "Type": "Float", + "Desc": "Lapalace smooth term in categorical feature spilt. Avoid the bias of small categories.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20 + ] + } + }, + { + "Name": "L2CategoricalRegularization", + "Type": "Float", + "Desc": "L2 Regularization for categorical split.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.5, + 1, + 5, + 10 + ] + } + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Sets the random seed for LightGBM to use.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelLightGBM" + }, + "Desc": "Parallel LightGBM Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRankingOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LightGbmRegressor", + "Desc": "LightGBM Regression", + "FriendlyName": "LightGBM Regressor", + "ShortName": "LightGBMR", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations.", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 50, + 100, + 150, + 200 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Shrinkage rate for trees, used to prevent over-fitting. Range: (0,1].", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "Maximum leaves for trees.", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of instances needed in a child.", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20, + 50 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "Booster", + "Type": { + "Kind": "Component", + "ComponentKind": "BoosterParameterFunction" + }, + "Desc": "Which booster to use, can be gbtree, gblinear or dart. gbtree and dart use tree based model while gblinear uses linear function.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": { + "Name": "gbdt" + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "EvaluationMetric", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Default", + "MeanAbsoluteError", + "RootMeanSquaredError", + "MeanSquaredError" + ] + }, + "Desc": "Evaluation metrics.", + "Aliases": [ + "em" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "RootMeanSquaredError" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of bucket bin for features.", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "Verbose", + "Type": "Bool", + "Desc": "Verbose", + "Aliases": [ + "v" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Silent", + "Type": "Bool", + "Desc": "Printing running messages.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of parallel threads used to run LightGBM.", + "Aliases": [ + "nt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "EarlyStoppingRound", + "Type": "Int", + "Desc": "Rounds of early stopping, 0 will disable it.", + "Aliases": [ + "es" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of entries in a batch when loading data.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1048576 + }, + { + "Name": "UseCategoricalSplit", + "Type": "Bool", + "Desc": "Enable categorical split or not.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "HandleMissingValue", + "Type": "Bool", + "Desc": "Enable special handling of missing value or not.", + "Aliases": [ + "hmv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "UseZeroAsMissingValue", + "Type": "Bool", + "Desc": "Enable usage of zero (0) as missing value.", + "Aliases": [ + "uzam" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "MinimumExampleCountPerGroup", + "Type": "Int", + "Desc": "Minimum number of instances per categorical group.", + "Aliases": [ + "mdpg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 50, + 100, + 200 + ] + } + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Max number of categorical thresholds.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 32, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 8, + 16, + 32, + 64 + ] + } + }, + { + "Name": "CategoricalSmoothing", + "Type": "Float", + "Desc": "Lapalace smooth term in categorical feature spilt. Avoid the bias of small categories.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20 + ] + } + }, + { + "Name": "L2CategoricalRegularization", + "Type": "Float", + "Desc": "L2 Regularization for categorical split.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.5, + 1, + 5, + 10 + ] + } + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Sets the random seed for LightGBM to use.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelLightGBM" + }, + "Desc": "Parallel LightGBM Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LinearSvmBinaryClassifier", + "Desc": "Train a linear SVM.", + "FriendlyName": "SVM (Pegasos-Linear)", + "ShortName": "svm", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight", + "WeightColumn" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Lambda", + "Type": "Float", + "Desc": "Regularizer constant", + "Aliases": [ + "lambda" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.001, + "SweepRange": { + "RangeType": "Float", + "Min": 1E-05, + "Max": 0.1, + "StepSize": 10.0, + "IsLogScale": true + } + }, + { + "Name": "PerformProjection", + "Type": "Bool", + "Desc": "Perform projection to unit-ball? Typically used with batch size > 1.", + "Aliases": [ + "project" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter", + "numIterations" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 100, + "StepSize": 10.0, + "IsLogScale": true + } + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "initWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "NoBias", + "Type": "Bool", + "Desc": "No bias", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "InitialWeights", + "Type": "String", + "Desc": "Initial Weights and bias, comma-separated", + "Aliases": [ + "initweights" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Whether to shuffle for each training iteration", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Batch size", + "Aliases": [ + "batch" + ], + "Required": false, + "SortOrder": 190.0, + "IsNullable": false, + "Default": 1 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LocalDeepSvmBinaryClassifier", + "Desc": "LD-SVM learns a binary, non-linear SVM classifier with a kernel that is specifically designed to reduce prediction time. LD-SVM learns decision boundaries that are locally linear.", + "FriendlyName": "Local Deep SVM (LDSVM)", + "ShortName": "LDSVM", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "TreeDepth", + "Type": "Int", + "Desc": "Depth of Local Deep SVM tree", + "Aliases": [ + "depth" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 3, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 3, + 5, + 7 + ] + } + }, + { + "Name": "LambdaW", + "Type": "Float", + "Desc": "Regularizer for classifier parameter W", + "Aliases": [ + "lw" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.1, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.01, + 0.001 + ] + } + }, + { + "Name": "LambdaTheta", + "Type": "Float", + "Desc": "Regularizer for kernel parameter Theta", + "Aliases": [ + "lt" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.01, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.01, + 0.001 + ] + } + }, + { + "Name": "LambdaThetaprime", + "Type": "Float", + "Desc": "Regularizer for kernel parameter Thetaprime", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.01, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.01, + 0.001 + ] + } + }, + { + "Name": "Sigma", + "Type": "Float", + "Desc": "Parameter for sigmoid sharpness", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1.0, + 0.1, + 0.01 + ] + } + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter", + "NumIterations" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 15000, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10000, + 15000 + ] + } + }, + { + "Name": "UseBias", + "Type": "Bool", + "Desc": "No bias", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "Cache", + "Type": "Bool", + "Desc": "Whether to cache the data before the first iteration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LogisticRegressionBinaryClassifier", + "Desc": "Logistic Regression is a method in statistics used to predict the probability of occurrence of an event and can be used as a classification algorithm. The algorithm predicts the probability of occurrence of an event by fitting data to a logistical function.", + "FriendlyName": "Logistic Regression", + "ShortName": "lr", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "ShowTrainingStatistics", + "Type": "Bool", + "Desc": "Show statistics of training examples.", + "Aliases": [ + "stat", + "ShowTrainingStats" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization weight", + "Aliases": [ + "l2", + "L2Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization weight", + "Aliases": [ + "l1", + "L1Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "OptimizationTolerance", + "Type": "Float", + "Desc": "Tolerance parameter for optimization convergence. Low = slower, more accurate", + "Aliases": [ + "ot", + "OptTol" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1E-07, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0001, + 1E-07 + ] + } + }, + { + "Name": "HistorySize", + "Type": "Int", + "Desc": "Memory size for L-BFGS. Low=faster, less accurate", + "Aliases": [ + "m", + "MemorySize" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 5, + 20, + 50 + ] + } + }, + { + "Name": "EnforceNonNegativity", + "Type": "Bool", + "Desc": "Enforce non-negative weights", + "Aliases": [ + "nn" + ], + "Required": false, + "SortOrder": 90.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "InitWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum iterations.", + "Aliases": [ + "maxiter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 2147483647 + } + }, + { + "Name": "StochasticGradientDescentInitilaizationTolerance", + "Type": "Float", + "Desc": "Run SGD to initialize LR weights, converging to this tolerance", + "Aliases": [ + "sgd", + "SgdInitializationTolerance" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Quiet", + "Type": "Bool", + "Desc": "If set to true, produce no output during training.", + "Aliases": [ + "q" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseThreads", + "Type": "Bool", + "Desc": "Whether or not to use threads. Default is true", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of threads", + "Aliases": [ + "nt", + "NumThreads" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DenseOptimizer", + "Type": "Bool", + "Desc": "Force densification of the internal optimization vectors", + "Aliases": [ + "do" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LogisticRegressionClassifier", + "Desc": "Maximum entropy classification is a method in statistics used to predict the probabilities of parallel events. The model predicts the probabilities of parallel events by fitting data to a softmax function.", + "FriendlyName": "Multi-class Logistic Regression", + "ShortName": "mlr", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "ShowTrainingStatistics", + "Type": "Bool", + "Desc": "Show statistics of training examples.", + "Aliases": [ + "stat", + "ShowTrainingStats" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization weight", + "Aliases": [ + "l2", + "L2Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization weight", + "Aliases": [ + "l1", + "L1Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "OptimizationTolerance", + "Type": "Float", + "Desc": "Tolerance parameter for optimization convergence. Low = slower, more accurate", + "Aliases": [ + "ot", + "OptTol" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1E-07, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0001, + 1E-07 + ] + } + }, + { + "Name": "HistorySize", + "Type": "Int", + "Desc": "Memory size for L-BFGS. Low=faster, less accurate", + "Aliases": [ + "m", + "MemorySize" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 5, + 20, + 50 + ] + } + }, + { + "Name": "EnforceNonNegativity", + "Type": "Bool", + "Desc": "Enforce non-negative weights", + "Aliases": [ + "nn" + ], + "Required": false, + "SortOrder": 90.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "InitWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum iterations.", + "Aliases": [ + "maxiter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 2147483647 + } + }, + { + "Name": "StochasticGradientDescentInitilaizationTolerance", + "Type": "Float", + "Desc": "Run SGD to initialize LR weights, converging to this tolerance", + "Aliases": [ + "sgd", + "SgdInitializationTolerance" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Quiet", + "Type": "Bool", + "Desc": "If set to true, produce no output during training.", + "Aliases": [ + "q" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseThreads", + "Type": "Bool", + "Desc": "Whether or not to use threads. Default is true", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of threads", + "Aliases": [ + "nt", + "NumThreads" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DenseOptimizer", + "Type": "Bool", + "Desc": "Force densification of the internal optimization vectors", + "Aliases": [ + "do" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.NaiveBayesClassifier", + "Desc": "Train a MulticlassNaiveBayesTrainer.", + "FriendlyName": "Multiclass Naive Bayes", + "ShortName": "MNB", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.OnlineGradientDescentRegressor", + "Desc": "Train a Online gradient descent perceptron.", + "FriendlyName": "Stochastic Gradient Descent (Regression)", + "ShortName": "ogd", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "RegressionLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "SquaredLoss" + } + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.1, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.01, + 0.1, + 0.5, + 1.0 + ] + } + }, + { + "Name": "DecreaseLearningRate", + "Type": "Bool", + "Desc": "Decrease learning rate", + "Aliases": [ + "decreaselr" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 Regularization Weight", + "Aliases": [ + "reg", + "L2RegularizerWeight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 0.4 + } + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter", + "numIterations" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 100, + "StepSize": 10.0, + "IsLogScale": true + } + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "initWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "ResetWeightsAfterXExamples", + "Type": "Int", + "Desc": "Number of examples after which weights will be reset to the current average", + "Aliases": [ + "numreset" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LazyUpdate", + "Type": "Bool", + "Desc": "Instead of updating averaged weights on every example, only update when loss is nonzero", + "Aliases": [ + "lazy", + "DoLazyUpdates" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "RecencyGain", + "Type": "Float", + "Desc": "Extra weight given to more recent updates", + "Aliases": [ + "rg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "RecencyGainMultiplicative", + "Type": "Bool", + "Desc": "Whether Recency Gain is multiplicative (vs. additive)", + "Aliases": [ + "rgm", + "RecencyGainMulti" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Averaged", + "Type": "Bool", + "Desc": "Do averaging?", + "Aliases": [ + "avg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "AveragedTolerance", + "Type": "Float", + "Desc": "The inexactness tolerance for averaging", + "Aliases": [ + "avgtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01 + }, + { + "Name": "InitialWeights", + "Type": "String", + "Desc": "Initial Weights and bias, comma-separated", + "Aliases": [ + "initweights" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Whether to shuffle for each training iteration", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.OrdinaryLeastSquaresRegressor", + "Desc": "Train an OLS regression model.", + "FriendlyName": "Ordinary Least Squares (Regression)", + "ShortName": "ols", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization weight", + "Aliases": [ + "l2" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1E-06, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1E-06, + 0.1, + 1.0 + ] + } + }, + { + "Name": "CalculateStatistics", + "Type": "Bool", + "Desc": "Whether to calculate per parameter significance statistics", + "Aliases": [ + "sig" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of entries in a batch when loading data (0 = auto).", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.PcaAnomalyDetector", + "Desc": "Train an PCA Anomaly model.", + "FriendlyName": "PCA Anomaly Detector", + "ShortName": "pcaAnom", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Rank", + "Type": "Int", + "Desc": "The number of components in the PCA", + "Aliases": [ + "k" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 40, + 80 + ] + } + }, + { + "Name": "Oversampling", + "Type": "Int", + "Desc": "Oversampling parameter for randomized PCA training", + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 40 + ] + } + }, + { + "Name": "Center", + "Type": "Bool", + "Desc": "If enabled, data is centered to be zero mean", + "Aliases": [ + "center" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed for random number generation", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "IUnsupervisedTrainerWithWeight", + "ITrainerInput" + ], + "OutputKind": [ + "IAnomalyDetectionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.PoissonRegressor", + "Desc": "Train an Poisson regression model.", + "FriendlyName": "Poisson Regression", + "ShortName": "PR", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization weight", + "Aliases": [ + "l2", + "L2Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization weight", + "Aliases": [ + "l1", + "L1Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "OptimizationTolerance", + "Type": "Float", + "Desc": "Tolerance parameter for optimization convergence. Low = slower, more accurate", + "Aliases": [ + "ot", + "OptTol" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1E-07, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0001, + 1E-07 + ] + } + }, + { + "Name": "HistorySize", + "Type": "Int", + "Desc": "Memory size for L-BFGS. Low=faster, less accurate", + "Aliases": [ + "m", + "MemorySize" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 5, + 20, + 50 + ] + } + }, + { + "Name": "EnforceNonNegativity", + "Type": "Bool", + "Desc": "Enforce non-negative weights", + "Aliases": [ + "nn" + ], + "Required": false, + "SortOrder": 90.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "InitWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum iterations.", + "Aliases": [ + "maxiter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 2147483647 + } + }, + { + "Name": "StochasticGradientDescentInitilaizationTolerance", + "Type": "Float", + "Desc": "Run SGD to initialize LR weights, converging to this tolerance", + "Aliases": [ + "sgd", + "SgdInitializationTolerance" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Quiet", + "Type": "Bool", + "Desc": "If set to true, produce no output during training.", + "Aliases": [ + "q" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseThreads", + "Type": "Bool", + "Desc": "Whether or not to use threads. Default is true", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of threads", + "Aliases": [ + "nt", + "NumThreads" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DenseOptimizer", + "Type": "Bool", + "Desc": "Force densification of the internal optimization vectors", + "Aliases": [ + "do" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.StochasticDualCoordinateAscentBinaryClassifier", + "Desc": "Train an SDCA binary model.", + "FriendlyName": "Fast Linear (SA-SDCA)", + "ShortName": "SDCA", + "Inputs": [ + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularizer constant. By default the l2 constant is automatically inferred based on data set.", + "Aliases": [ + "l2", + "L2Const" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 1E-07, + 1E-06, + 1E-05, + 0.0001, + 0.001, + 0.01 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "L1Threshold", + "Type": "Float", + "Desc": "L1 soft threshold (L1/L2). Note that it is easier to control and sweep using the threshold parameter than the raw L1-regularizer constant. By default the l1 threshold is automatically inferred based on data set.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 0.0, + 0.25, + 0.5, + 0.75, + 1.0 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "SDCAClassificationLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "LogLoss" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Defaults to automatic. Determinism not guaranteed.", + "Aliases": [ + "nt", + "t", + "threads", + "NumThreads" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "PositiveInstanceWeight", + "Type": "Float", + "Desc": "Apply weight to the positive class, for imbalanced data", + "Aliases": [ + "piw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "ConvergenceTolerance", + "Type": "Float", + "Desc": "The tolerance for the ratio between duality gap and primal loss for convergence checking.", + "Aliases": [ + "tol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.001, + 0.01, + 0.1, + 0.2 + ] + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum number of iterations; set to 1 to simulate online learning. Defaults to automatic.", + "Aliases": [ + "iter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 10, + 20, + 100 + ] + } + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Shuffle data every epoch?", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "ConvergenceCheckFrequency", + "Type": "Int", + "Desc": "Convergence check frequency (in terms of number of iterations). Set as negative or zero for not checking at all. If left blank, it defaults to check after every 'numThreads' iterations.", + "Aliases": [ + "checkFreq", + "CheckFrequency" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "BiasLearningRate", + "Type": "Float", + "Desc": "The learning rate for adjusting bias from being regularized.", + "Aliases": [ + "blr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.01, + 0.1, + 1.0 + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.StochasticDualCoordinateAscentClassifier", + "Desc": "The SDCA linear multi-class classification trainer.", + "FriendlyName": "Fast Linear Multi-class Classification (SA-SDCA)", + "ShortName": "sasdcamc", + "Inputs": [ + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularizer constant. By default the l2 constant is automatically inferred based on data set.", + "Aliases": [ + "l2", + "L2Const" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 1E-07, + 1E-06, + 1E-05, + 0.0001, + 0.001, + 0.01 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "L1Threshold", + "Type": "Float", + "Desc": "L1 soft threshold (L1/L2). Note that it is easier to control and sweep using the threshold parameter than the raw L1-regularizer constant. By default the l1 threshold is automatically inferred based on data set.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 0.0, + 0.25, + 0.5, + 0.75, + 1.0 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "SDCAClassificationLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "LogLoss" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Defaults to automatic. Determinism not guaranteed.", + "Aliases": [ + "nt", + "t", + "threads", + "NumThreads" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ConvergenceTolerance", + "Type": "Float", + "Desc": "The tolerance for the ratio between duality gap and primal loss for convergence checking.", + "Aliases": [ + "tol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.001, + 0.01, + 0.1, + 0.2 + ] + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum number of iterations; set to 1 to simulate online learning. Defaults to automatic.", + "Aliases": [ + "iter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 10, + 20, + 100 + ] + } + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Shuffle data every epoch?", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "ConvergenceCheckFrequency", + "Type": "Int", + "Desc": "Convergence check frequency (in terms of number of iterations). Set as negative or zero for not checking at all. If left blank, it defaults to check after every 'numThreads' iterations.", + "Aliases": [ + "checkFreq", + "CheckFrequency" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "BiasLearningRate", + "Type": "Float", + "Desc": "The learning rate for adjusting bias from being regularized.", + "Aliases": [ + "blr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.01, + 0.1, + 1.0 + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.StochasticDualCoordinateAscentRegressor", + "Desc": "The SDCA linear regression trainer.", + "FriendlyName": "Fast Linear Regression (SA-SDCA)", + "ShortName": "sasdcar", + "Inputs": [ + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularizer constant. By default the l2 constant is automatically inferred based on data set.", + "Aliases": [ + "l2", + "L2Const" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 1E-07, + 1E-06, + 1E-05, + 0.0001, + 0.001, + 0.01 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "L1Threshold", + "Type": "Float", + "Desc": "L1 soft threshold (L1/L2). Note that it is easier to control and sweep using the threshold parameter than the raw L1-regularizer constant. By default the l1 threshold is automatically inferred based on data set.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 0.0, + 0.25, + 0.5, + 0.75, + 1.0 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "SDCARegressionLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "SquaredLoss" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Defaults to automatic. Determinism not guaranteed.", + "Aliases": [ + "nt", + "t", + "threads", + "NumThreads" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ConvergenceTolerance", + "Type": "Float", + "Desc": "The tolerance for the ratio between duality gap and primal loss for convergence checking.", + "Aliases": [ + "tol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.001, + 0.01, + 0.1, + 0.2 + ] + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum number of iterations; set to 1 to simulate online learning. Defaults to automatic.", + "Aliases": [ + "iter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 10, + 20, + 100 + ] + } + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Shuffle data every epoch?", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "ConvergenceCheckFrequency", + "Type": "Int", + "Desc": "Convergence check frequency (in terms of number of iterations). Set as negative or zero for not checking at all. If left blank, it defaults to check after every 'numThreads' iterations.", + "Aliases": [ + "checkFreq", + "CheckFrequency" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "BiasLearningRate", + "Type": "Float", + "Desc": "The learning rate for adjusting bias from being regularized.", + "Aliases": [ + "blr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.01, + 0.1, + 1.0 + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.StochasticGradientDescentBinaryClassifier", + "Desc": "Train an Hogwild SGD binary model.", + "FriendlyName": "Hogwild SGD (binary)", + "ShortName": "HogwildSGD", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "ClassificationLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "LogLoss" + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 Regularization constant", + "Aliases": [ + "l2", + "L2Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1E-06, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1E-07, + 5E-07, + 1E-06, + 5E-06, + 1E-05 + ] + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Defaults to automatic depending on data sparseness. Determinism not guaranteed.", + "Aliases": [ + "nt", + "t", + "threads", + "NumThreads" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "ConvergenceTolerance", + "Type": "Float", + "Desc": "Exponential moving averaged improvement tolerance for convergence", + "Aliases": [ + "tol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0001, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.01, + 0.001, + 0.0001, + 1E-05 + ] + } + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Maximum number of iterations; set to 1 to simulate online learning.", + "Aliases": [ + "iter", + "MaxIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 5, + 10, + 20 + ] + } + }, + { + "Name": "InitialLearningRate", + "Type": "Float", + "Desc": "Initial learning rate (only used by SGD)", + "Aliases": [ + "ilr", + "lr", + "InitLearningRate" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01 + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Shuffle data every epoch?", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "PositiveInstanceWeight", + "Type": "Float", + "Desc": "Apply weight to the positive class, for imbalanced data", + "Aliases": [ + "piw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "CheckFrequency", + "Type": "Int", + "Desc": "Convergence check frequency (in terms of number of iterations). Default equals number of threads", + "Aliases": [ + "checkFreq" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.SymSgdBinaryClassifier", + "Desc": "Train a symbolic SGD.", + "FriendlyName": "Symbolic SGD (binary)", + "ShortName": "SymSGD", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of passes over the data.", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 50, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 5, + 10, + 20, + 30, + 40, + 50 + ] + } + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 51.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 10.0, + 1.0, + 0.1, + 0.01, + 0.001 + ] + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization", + "Aliases": [ + "l2" + ], + "Required": false, + "SortOrder": 52.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-05, + 1E-05, + 1E-06, + 1E-07 + ] + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Determinism not guaranteed. Multi-threading is not supported currently.", + "Aliases": [ + "nt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Tolerance", + "Type": "Float", + "Desc": "Tolerance for difference in average loss in consecutive passes.", + "Aliases": [ + "tol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0001 + }, + { + "Name": "UpdateFrequency", + "Type": "Int", + "Desc": "The number of iterations each thread learns a local model until combining it with the global model. Low value means more updated global model and high value means less cache traffic.", + "Aliases": [ + "freq" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 5, + 20 + ] + } + }, + { + "Name": "MemorySize", + "Type": "Int", + "Desc": "The acceleration memory budget in MB", + "Aliases": [ + "accelMemBudget" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1024 + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Shuffle data?", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "PositiveInstanceWeight", + "Type": "Float", + "Desc": "Apply weight to the positive class, for imbalanced data", + "Aliases": [ + "piw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Transforms.ApproximateBootstrapSampler", + "Desc": "Approximate bootstrap sampling.", + "FriendlyName": "Bootstrap Sample Transform", + "ShortName": "BootstrapSample", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Complement", + "Type": "Bool", + "Desc": "Whether this is the out-of-bag sample, that is, all those rows that are not selected by the transform.", + "Aliases": [ + "comp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "The random seed. If unspecified random state will be instead derived from the environment.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ShuffleInput", + "Type": "Bool", + "Desc": "Whether we should attempt to shuffle the source data. By default on, but can be turned off for efficiency.", + "Aliases": [ + "si" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "PoolSize", + "Type": "Int", + "Desc": "When shuffling the output, the number of output rows to keep in that pool. Note that shuffling of output is completely distinct from shuffling of input.", + "Aliases": [ + "pool" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.BinaryPredictionScoreColumnsRenamer", + "Desc": "For binary prediction, it renames the PredictedLabel and Score columns to include the name of the positive class.", + "FriendlyName": "Rename Binary Prediction Score Columns", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model used in scoring", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.BinNormalizer", + "Desc": "The values are assigned into equidensity bins and a value is mapped to its bin_number/number_of_bins.", + "FriendlyName": "Binning Normalizer", + "ShortName": "Bin", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "NumBins", + "Type": "Int", + "Desc": "Max number of bins, power of 2 recommended", + "Aliases": [ + "bins" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumBins", + "Type": "Int", + "Desc": "Max number of bins, power of 2 recommended", + "Aliases": [ + "bins" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1024 + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.CategoricalHashOneHotVectorizer", + "Desc": "Converts the categorical value into an indicator array by hashing the value and using the hash as an index in the bag. If the input column is a vector, a single indicator bag is returned for it.", + "FriendlyName": "Categorical Hash Transform", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "OutputKind", + "Type": { + "Kind": "Enum", + "Values": [ + "Bag", + "Indicator", + "Key", + "Binary" + ] + }, + "Desc": "Output kind: Bag (multi-set vector), Ind (indicator vector), or Key (index)", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "The number of bits to hash into. Must be between 1 and 30, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Ordered", + "Type": "Bool", + "Desc": "Whether the position of each term should be included in the hash", + "Aliases": [ + "ord" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaximumNumberOfInverts", + "Type": "Int", + "Desc": "Limit the number of keys used to generate the slot name to this many. 0 means no invert hashing, -1 means no limit.", + "Aliases": [ + "ih" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:numberOfBits:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "Number of bits to hash into. Must be between 1 and 30, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 16 + }, + { + "Name": "OutputKind", + "Type": { + "Kind": "Enum", + "Values": [ + "Bag", + "Indicator", + "Key", + "Binary" + ] + }, + "Desc": "Output kind: Bag (multi-set vector), Ind (indicator vector), or Key (index)", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": "Bag" + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + }, + { + "Name": "Ordered", + "Type": "Bool", + "Desc": "Whether the position of each term should be included in the hash", + "Aliases": [ + "ord" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaximumNumberOfInverts", + "Type": "Int", + "Desc": "Limit the number of keys used to generate the slot name to this many. 0 means no invert hashing, -1 means no limit.", + "Aliases": [ + "ih" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.CategoricalOneHotVectorizer", + "Desc": "Converts the categorical value into an indicator array by building a dictionary of categories based on the data and using the id in the dictionary as the index in the array.", + "FriendlyName": "Categorical Transform", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "OutputKind", + "Type": { + "Kind": "Enum", + "Values": [ + "Bag", + "Indicator", + "Key", + "Binary" + ] + }, + "Desc": "Output kind: Bag (multi-set vector), Ind (indicator vector), Key (index), or Binary encoded indicator vector", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of terms to keep when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of keys to keep per column when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "OutputKind", + "Type": { + "Kind": "Enum", + "Values": [ + "Bag", + "Indicator", + "Key", + "Binary" + ] + }, + "Desc": "Output kind: Bag (multi-set vector), Ind (indicator vector), or Key (index)", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": "Indicator" + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 113.0, + "IsNullable": false, + "Default": "ByOccurrence" + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 114.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.CharacterTokenizer", + "Desc": "Character-oriented tokenizer where text is considered a sequence of characters.", + "FriendlyName": "Character Tokenizer Transform", + "ShortName": "CharToken", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseMarkerChars", + "Type": "Bool", + "Desc": "Whether to mark the beginning/end of each row/slot with start of text character (0x02)/end of text character (0x03)", + "Aliases": [ + "mark" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ColumnConcatenator", + "Desc": "Concatenates one or more columns of the same item type.", + "FriendlyName": "Concat Transform", + "ShortName": "Concat", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:srcs)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ColumnCopier", + "Desc": "Duplicates columns from the dataset", + "FriendlyName": "Copy Columns Transform", + "ShortName": "Copy", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ColumnSelector", + "Desc": "Selects a set of columns, dropping all others", + "FriendlyName": "Select Columns", + "ShortName": null, + "Inputs": [ + { + "Name": "KeepColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of columns to keep.", + "Aliases": [ + "keepcol" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "DropColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of columns to drop.", + "Aliases": [ + "dropcol" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "KeepHidden", + "Type": "Bool", + "Desc": "Specifies whether to keep or remove hidden columns.", + "Aliases": [ + "hidden" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "IgnoreMissing", + "Type": "Bool", + "Desc": "Specifies whether to ignore columns that are missing from the input.", + "Aliases": [ + "ignore" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ColumnTypeConverter", + "Desc": "Converts a column to a different type, using standard conversions.", + "FriendlyName": "Convert Transform", + "ShortName": "Convert", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "ResultType", + "Type": { + "Kind": "Enum", + "Values": [ + "I1", + "U1", + "I2", + "U2", + "I4", + "U4", + "I8", + "U8", + "R4", + "Num", + "R8", + "TX", + "Text", + "TXT", + "BL", + "Bool", + "TimeSpan", + "TS", + "DT", + "DateTime", + "DZ", + "DateTimeZone", + "UG", + "U16" + ] + }, + "Desc": "The result type", + "Aliases": [ + "type" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Range", + "Type": "String", + "Desc": "For a key column, this defines the range of values", + "Aliases": [ + "key" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:type:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ResultType", + "Type": { + "Kind": "Enum", + "Values": [ + "I1", + "U1", + "I2", + "U2", + "I4", + "U4", + "I8", + "U8", + "R4", + "Num", + "R8", + "TX", + "Text", + "TXT", + "BL", + "Bool", + "TimeSpan", + "TS", + "DT", + "DateTime", + "DZ", + "DateTimeZone", + "UG", + "U16" + ] + }, + "Desc": "The result type", + "Aliases": [ + "type" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Range", + "Type": "String", + "Desc": "For a key column, this defines the range of values", + "Aliases": [ + "key" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.CombinerByContiguousGroupId", + "Desc": "Groups values of a scalar column into a vector, by a contiguous group ID", + "FriendlyName": "Group Transform", + "ShortName": "Group", + "Inputs": [ + { + "Name": "GroupKey", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Columns to group by", + "Aliases": [ + "g" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Columns to group together", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ConditionalNormalizer", + "Desc": "Normalize the columns only if needed", + "FriendlyName": "Normalize If Needed", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ] + }, + { + "Name": "Transforms.CountTableBuilder", + "Desc": "Transforms the categorical column into the set of features: count of each label class, log-odds for each label class, back-off indicator. The input columns must be keys.", + "FriendlyName": "Count Table Transform", + "ShortName": "Count", + "Inputs": [ + { + "Name": "Columns", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "CountTable", + "Type": { + "Kind": "Component", + "ComponentKind": "CountTableBuilder" + }, + "Desc": "Count table settings", + "Aliases": [ + "table" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PriorCoefficient", + "Type": "Float", + "Desc": "The coefficient with which to apply the prior smoothing to the features", + "Aliases": [ + "prior" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LaplaceScale", + "Type": "Float", + "Desc": "Laplacian noise diversity/scale-parameter. Suggest keeping it less than 1.", + "Aliases": [ + "laplace" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Seed for the random generator for the laplacian noise.", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "CountTable", + "Type": { + "Kind": "Component", + "ComponentKind": "CountTableBuilder" + }, + "Desc": "Count table settings", + "Aliases": [ + "table" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "CMSketch" + } + }, + { + "Name": "PriorCoefficient", + "Type": "Float", + "Desc": "The coefficient with which to apply the prior smoothing to the features", + "Aliases": [ + "prior" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "LaplaceScale", + "Type": "Float", + "Desc": "Laplacian noise diversity/scale-parameter. Suggest keeping it less than 1.", + "Aliases": [ + "laplace" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Seed for the random generator for the laplacian noise.", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Label column", + "Aliases": [ + "label", + "lab" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + }, + { + "Name": "InitialCountsModel", + "Type": "String", + "Desc": "Optional model file to load counts from. If this is specified all other options are ignored.", + "Aliases": [ + "inmodel", + "extfile" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "SharedTable", + "Type": "Bool", + "Desc": "Keep counts for all columns in one shared count table", + "Aliases": [ + "shared" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.CountTargetEncoder", + "Desc": "Transforms the categorical column into the set of features: count of each label class, log-odds for each label class, back-off indicator. The columns can be of arbitrary type.", + "FriendlyName": "Count Table Transform", + "ShortName": "Count", + "Inputs": [ + { + "Name": "Columns", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "CountTable", + "Type": { + "Kind": "Component", + "ComponentKind": "CountTableBuilder" + }, + "Desc": "Count table settings", + "Aliases": [ + "table" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PriorCoefficient", + "Type": "Float", + "Desc": "The coefficient with which to apply the prior smoothing to the features", + "Aliases": [ + "prior" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LaplaceScale", + "Type": "Float", + "Desc": "Laplacian noise diversity/scale-parameter. Suggest keeping it less than 1.", + "Aliases": [ + "laplace" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Seed for the random generator for the laplacian noise.", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Combine", + "Type": "Bool", + "Desc": "Whether the values need to be combined for a single hash", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Combine", + "Type": "Bool", + "Desc": "Whether the values need to be combined for a single hash", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CountTable", + "Type": { + "Kind": "Component", + "ComponentKind": "CountTableBuilder" + }, + "Desc": "Count table settings", + "Aliases": [ + "table" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "CMSketch" + } + }, + { + "Name": "PriorCoefficient", + "Type": "Float", + "Desc": "The coefficient with which to apply the prior smoothing to the features", + "Aliases": [ + "prior" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "LaplaceScale", + "Type": "Float", + "Desc": "Laplacian noise diversity/scale-parameter. Suggest keeping it less than 1.", + "Aliases": [ + "laplace" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Seed for the random generator for the laplacian noise.", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Label column", + "Aliases": [ + "label", + "lab" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + }, + { + "Name": "InitialCountsModel", + "Type": "String", + "Desc": "Optional model file to load counts from. If this is specified all other options are ignored.", + "Aliases": [ + "inmodel", + "extfile" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "SharedTable", + "Type": "Bool", + "Desc": "Keep counts for all columns in one shared count table", + "Aliases": [ + "shared" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "Number of bits to hash into. Must be between 1 and 31, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 31 + }, + { + "Name": "HashingSeed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.DatasetScorer", + "Desc": "Score a dataset with a predictor model", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The dataset to be scored", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model to apply to data", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Suffix", + "Type": "String", + "Desc": "Suffix to append to the score columns", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "ScoredData", + "Type": "DataView", + "Desc": "The scored dataset" + }, + { + "Name": "ScoringTransform", + "Type": "TransformModel", + "Desc": "The scoring transform" + } + ] + }, + { + "Name": "Transforms.DatasetTransformScorer", + "Desc": "Score a dataset with a transform model", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The dataset to be scored", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "TransformModel", + "Type": "TransformModel", + "Desc": "The transform model to apply to data", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "ScoredData", + "Type": "DataView", + "Desc": "The scored dataset" + }, + { + "Name": "ScoringTransform", + "Type": "TransformModel", + "Desc": "The scoring transform" + } + ] + }, + { + "Name": "Transforms.Dictionarizer", + "Desc": "Converts input values (words, numbers, etc.) to index in a dictionary.", + "FriendlyName": "Term Transform", + "ShortName": "TermTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of terms to keep when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of keys to keep per column when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 113.0, + "IsNullable": false, + "Default": "ByOccurrence" + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 114.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.FeatureCombiner", + "Desc": "Combines all the features into one feature column.", + "FriendlyName": "Feature Combiner", + "ShortName": "fc", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Features", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Features", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.FeatureContributionCalculationTransformer", + "Desc": "For each data point, calculates the contribution of individual features to the model prediction.", + "FriendlyName": "Feature Contribution Calculation", + "ShortName": null, + "Inputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model to apply to data", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumn", + "Type": "String", + "Desc": "Name of feature column", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "Top", + "Type": "Int", + "Desc": "Number of top contributions", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "Bottom", + "Type": "Int", + "Desc": "Number of bottom contributions", + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "Normalize", + "Type": "Bool", + "Desc": "Whether or not output of Features contribution should be normalized", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.FeatureSelectorByCount", + "Desc": "Selects the slots for which the count of non-default values is greater than or equal to a threshold.", + "FriendlyName": "Count Feature Selection Transform", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Columns to use for feature selection", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Count", + "Type": "Int", + "Desc": "If the count of non-default values for a slot is greater than or equal to this threshold, the slot is preserved", + "Aliases": [ + "c" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.FeatureSelectorByMutualInformation", + "Desc": "Selects the top k slots across all specified columns ordered by their mutual information with the label column.", + "FriendlyName": "Mutual Information Feature Selection Transform", + "ShortName": "MIFeatureSelection", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Columns to use for feature selection", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "SlotsInOutput", + "Type": "Int", + "Desc": "The maximum number of slots to preserve in output", + "Aliases": [ + "topk", + "numSlotsToKeep" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 1000 + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "NumBins", + "Type": "Int", + "Desc": "Max number of bins for R4/R8 columns, power of 2 recommended", + "Aliases": [ + "bins" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 256 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.GlobalContrastNormalizer", + "Desc": "Performs a global contrast normalization on input values: Y = (s * X - M) / D, where s is a scale, M is mean and D is either L2 norm or standard deviation.", + "FriendlyName": "Global Contrast Normalization Transform", + "ShortName": "Gcn", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "UseStdDev", + "Type": "Bool", + "Desc": "Normalize by standard deviation rather than L2 norm", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale features by this value", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "SubMean", + "Type": "Bool", + "Desc": "Subtract mean from each value before normalizing", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "SubMean", + "Type": "Bool", + "Desc": "Subtract mean from each value before normalizing", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseStdDev", + "Type": "Bool", + "Desc": "Normalize by standard deviation rather than L2 norm", + "Aliases": [ + "useStd" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale features by this value", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.HashConverter", + "Desc": "Converts column values into hashes. This transform accepts both numeric and text inputs, both single and vector-valued columns. ", + "FriendlyName": "Hash Join Transform", + "ShortName": "HashJoin", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Join", + "Type": "Bool", + "Desc": "Whether the values need to be combined for a single hash", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "CustomSlotMap", + "Type": "String", + "Desc": "Which slots should be combined together. Example: 0,3,5;0,1;3;2,1,0. Overrides 'join'.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "Number of bits to hash into. Must be between 1 and 31, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Ordered", + "Type": "Bool", + "Desc": "Whether the position of each term should be included in the hash", + "Aliases": [ + "ord" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "Number of bits to hash into. Must be between 1 and 31, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 31 + }, + { + "Name": "Join", + "Type": "Bool", + "Desc": "Whether the values need to be combined for a single hash", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + }, + { + "Name": "Ordered", + "Type": "Bool", + "Desc": "Whether the position of each term should be included in the hash", + "Aliases": [ + "ord" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ImageGrayscale", + "Desc": "Convert image into grayscale.", + "FriendlyName": "Image Greyscale Transform", + "ShortName": "ImageGrayscaleTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ImageLoader", + "Desc": "Load images from files.", + "FriendlyName": "Image Loader Transform", + "ShortName": "ImageLoaderTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ImageFolder", + "Type": "String", + "Desc": "Folder where to search for images", + "Aliases": [ + "folder" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ImagePixelExtractor", + "Desc": "Extract color plane(s) from an image. Options include scaling, offset and conversion to floating point.", + "FriendlyName": "Image Pixel Extractor Transform", + "ShortName": "ImagePixelExtractor", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "UseAlpha", + "Type": "Bool", + "Desc": "Whether to use alpha channel", + "Aliases": [ + "alpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "UseRed", + "Type": "Bool", + "Desc": "Whether to use red channel", + "Aliases": [ + "red" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "UseGreen", + "Type": "Bool", + "Desc": "Whether to use green channel", + "Aliases": [ + "green" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "UseBlue", + "Type": "Bool", + "Desc": "Whether to use blue channel", + "Aliases": [ + "blue" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Order", + "Type": { + "Kind": "Enum", + "Values": [ + "ARGB", + "ARBG", + "ABRG", + "ABGR", + "AGRB", + "AGBR" + ] + }, + "Desc": "Order of channels", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Interleave", + "Type": "Bool", + "Desc": "Whether to separate each channel or interleave in specified order", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Convert", + "Type": "Bool", + "Desc": "Whether to convert to floating point", + "Aliases": [ + "conv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "Offset (pre-scale)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale factor", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseAlpha", + "Type": "Bool", + "Desc": "Whether to use alpha channel", + "Aliases": [ + "alpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseRed", + "Type": "Bool", + "Desc": "Whether to use red channel", + "Aliases": [ + "red" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "UseGreen", + "Type": "Bool", + "Desc": "Whether to use green channel", + "Aliases": [ + "green" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "UseBlue", + "Type": "Bool", + "Desc": "Whether to use blue channel", + "Aliases": [ + "blue" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Order", + "Type": { + "Kind": "Enum", + "Values": [ + "ARGB", + "ARBG", + "ABRG", + "ABGR", + "AGRB", + "AGBR" + ] + }, + "Desc": "Order of colors.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "ARGB" + }, + { + "Name": "Interleave", + "Type": "Bool", + "Desc": "Whether to separate each channel or interleave in specified order", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Convert", + "Type": "Bool", + "Desc": "Whether to convert to floating point", + "Aliases": [ + "conv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "Offset (pre-scale)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale factor", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ImageResizer", + "Desc": "Scales an image to specified dimensions using one of the three scale types: isotropic with padding, isotropic with cropping or anisotropic. In case of isotropic padding, transparent color is used to pad resulting image.", + "FriendlyName": "Image Resizer Transform", + "ShortName": "ImageScalerTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "ImageWidth", + "Type": "Int", + "Desc": "Width of the resized image", + "Aliases": [ + "width" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ImageHeight", + "Type": "Int", + "Desc": "Height of the resized image", + "Aliases": [ + "height" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Resizing", + "Type": { + "Kind": "Enum", + "Values": [ + "IsoPad", + "IsoCrop", + "Fill" + ] + }, + "Desc": "Resizing method", + "Aliases": [ + "scale" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "CropAnchor", + "Type": { + "Kind": "Enum", + "Values": [ + "Right", + "Left", + "Top", + "Bottom", + "Center" + ] + }, + "Desc": "Anchor for cropping", + "Aliases": [ + "anchor" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ImageWidth", + "Type": "Int", + "Desc": "Resized width of the image", + "Aliases": [ + "width" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "ImageHeight", + "Type": "Int", + "Desc": "Resized height of the image", + "Aliases": [ + "height" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Resizing", + "Type": { + "Kind": "Enum", + "Values": [ + "IsoPad", + "IsoCrop", + "Fill" + ] + }, + "Desc": "Resizing method", + "Aliases": [ + "scale" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "IsoCrop" + }, + { + "Name": "CropAnchor", + "Type": { + "Kind": "Enum", + "Values": [ + "Right", + "Left", + "Top", + "Bottom", + "Center" + ] + }, + "Desc": "Anchor for cropping", + "Aliases": [ + "anchor" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Center" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.KeyToTextConverter", + "Desc": "KeyToValueTransform utilizes KeyValues metadata to map key indices to the corresponding values in the KeyValues metadata.", + "FriendlyName": "Key To Value Transform", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LabelColumnKeyBooleanConverter", + "Desc": "Transforms the label to either key or bool (if needed) to make it suitable for classification.", + "FriendlyName": "Prepare Classification Label", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "The label column", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Convert the key values to text", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LabelIndicator", + "Desc": "Label remapper used by OVA", + "FriendlyName": "LabelIndicator", + "ShortName": "LabelIndictator", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "ClassIndex", + "Type": "Int", + "Desc": "The positive example class for binary classification.", + "Aliases": [ + "index" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ClassIndex", + "Type": "Int", + "Desc": "Label of the positive class.", + "Aliases": [ + "index" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LabelToFloatConverter", + "Desc": "Transforms the label to float to make it suitable for regression.", + "FriendlyName": "Prepare Regression Label", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "The label column", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LightLda", + "Desc": "The LDA transform implements LightLDA, a state-of-the-art implementation of Latent Dirichlet Allocation.", + "FriendlyName": "Latent Dirichlet Allocation Transform", + "ShortName": "LightLda", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "NumTopic", + "Type": "Int", + "Desc": "The number of topics", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "AlphaSum", + "Type": "Float", + "Desc": "Dirichlet prior on document-topic vectors", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Beta", + "Type": "Float", + "Desc": "Dirichlet prior on vocab-topic vectors", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Mhstep", + "Type": "Int", + "Desc": "Number of Metropolis Hasting step", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LikelihoodInterval", + "Type": "Int", + "Desc": "Compute log likelihood over local dataset on this iteration interval", + "Aliases": [ + "llInterval" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumThreads", + "Type": "Int", + "Desc": "The number of training threads", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumMaxDocToken", + "Type": "Int", + "Desc": "The threshold of maximum count of tokens per doc", + "Aliases": [ + "maxNumToken" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumSummaryTermPerTopic", + "Type": "Int", + "Desc": "The number of words to summarize the topic", + "Aliases": [ + "ns" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumBurninIterations", + "Type": "Int", + "Desc": "The number of burn-in iterations", + "Aliases": [ + "burninIter" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": 10 + }, + { + "Name": "ResetRandomGenerator", + "Type": "Bool", + "Desc": "Reset the random number generator for each document", + "Aliases": [ + "reset" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "OutputTopicWordSummary", + "Type": "Bool", + "Desc": "Whether to output the topic-word summary in text format when saving the model to disk", + "Aliases": [ + "summary" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:srcs)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 49.0, + "IsNullable": false + }, + { + "Name": "NumTopic", + "Type": "Int", + "Desc": "The number of topics", + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 40, + 100, + 200 + ] + } + }, + { + "Name": "NumThreads", + "Type": "Int", + "Desc": "The number of training threads. Default value depends on number of logical processors.", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "NumMaxDocToken", + "Type": "Int", + "Desc": "The threshold of maximum count of tokens per doc", + "Aliases": [ + "maxNumToken" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 512 + }, + { + "Name": "AlphaSum", + "Type": "Float", + "Desc": "Dirichlet prior on document-topic vectors", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 100, + 200 + ] + } + }, + { + "Name": "Beta", + "Type": "Float", + "Desc": "Dirichlet prior on vocab-topic vectors", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.01, + 0.015, + 0.07, + 0.02 + ] + } + }, + { + "Name": "Mhstep", + "Type": "Int", + "Desc": "Number of Metropolis Hasting step", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 4, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 2, + 4, + 8, + 16 + ] + } + }, + { + "Name": "NumIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 200, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 100, + 200, + 300, + 400 + ] + } + }, + { + "Name": "LikelihoodInterval", + "Type": "Int", + "Desc": "Compute log likelihood over local dataset on this iteration interval", + "Aliases": [ + "llInterval" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "NumSummaryTermPerTopic", + "Type": "Int", + "Desc": "The number of words to summarize the topic", + "Aliases": [ + "ns" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "NumBurninIterations", + "Type": "Int", + "Desc": "The number of burn-in iterations", + "Aliases": [ + "burninIter" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 30, + 40 + ] + } + }, + { + "Name": "ResetRandomGenerator", + "Type": "Bool", + "Desc": "Reset the random number generator for each document", + "Aliases": [ + "reset" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "OutputTopicWordSummary", + "Type": "Bool", + "Desc": "Whether to output the topic-word summary in text format when saving the model to disk", + "Aliases": [ + "summary" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LogMeanVarianceNormalizer", + "Desc": "Normalizes the data based on the computed mean and variance of the logarithm of the data.", + "FriendlyName": "LogMeanVar Normalizer", + "ShortName": "LogMeanVar", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseCdf", + "Type": "Bool", + "Desc": "Whether to use CDF as the output", + "Aliases": [ + "cdf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LpNormalizer", + "Desc": "Normalize vectors (rows) individually by rescaling them to unit norm (L2, L1 or LInf). Performs the following operation on a vector X: Y = (X - M) / D, where M is mean and D is either L2 norm, L1 norm or LInf norm.", + "FriendlyName": "Lp-Norm Normalizer", + "ShortName": "lpnorm", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Norm", + "Type": { + "Kind": "Enum", + "Values": [ + "L2", + "StandardDeviation", + "L1", + "Infinity" + ] + }, + "Desc": "The norm to use to normalize each sample", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "SubMean", + "Type": "Bool", + "Desc": "Subtract mean from each value before normalizing", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Norm", + "Type": { + "Kind": "Enum", + "Values": [ + "L2", + "StandardDeviation", + "L1", + "Infinity" + ] + }, + "Desc": "The norm to use to normalize each sample", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": "L2" + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "SubMean", + "Type": "Bool", + "Desc": "Subtract mean from each value before normalizing", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ManyHeterogeneousModelCombiner", + "Desc": "Combines a sequence of TransformModels and a PredictorModel into a single PredictorModel.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "TransformModels", + "Type": { + "Kind": "Array", + "ItemType": "TransformModel" + }, + "Desc": "Transform model", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Predictor model", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Predictor model" + } + ] + }, + { + "Name": "Transforms.MeanVarianceNormalizer", + "Desc": "Normalizes the data based on the computed mean and variance of the data.", + "FriendlyName": "MeanVar Normalizer", + "ShortName": "MeanVar", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseCdf", + "Type": "Bool", + "Desc": "Whether to use CDF as the output", + "Aliases": [ + "cdf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MinMaxNormalizer", + "Desc": "Normalizes the data based on the observed minimum and maximum values of the data.", + "FriendlyName": "Min-Max Normalizer", + "ShortName": "MinMax", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MissingValueHandler", + "Desc": "Handle missing values by replacing them with either the default value or the mean/min/max value (for non-text columns only). An indicator column can optionally be concatenated, if the input column type is numeric.", + "FriendlyName": "NA Handle Transform", + "ShortName": "NAHandle", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Kind", + "Type": { + "Kind": "Enum", + "Values": [ + "DefaultValue", + "Mean", + "Minimum", + "Maximum" + ] + }, + "Desc": "The replacement method to utilize", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ImputeBySlot", + "Type": "Bool", + "Desc": "Whether to impute values by slot", + "Aliases": [ + "slot" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ConcatIndicator", + "Type": "Bool", + "Desc": "Whether or not to concatenate an indicator vector column to the value column", + "Aliases": [ + "ind" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:rep:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ReplaceWith", + "Type": { + "Kind": "Enum", + "Values": [ + "DefaultValue", + "Mean", + "Minimum", + "Maximum" + ] + }, + "Desc": "The replacement method to utilize", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Def" + }, + { + "Name": "ImputeBySlot", + "Type": "Bool", + "Desc": "Whether to impute values by slot", + "Aliases": [ + "slot" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Concat", + "Type": "Bool", + "Desc": "Whether or not to concatenate an indicator vector column to the value column", + "Aliases": [ + "ind" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MissingValueIndicator", + "Desc": "Create a boolean output column with the same number of slots as the input column, where the output value is true if the value in the input column is missing.", + "FriendlyName": "NA Indicator Transform", + "ShortName": "NAInd", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MissingValuesDropper", + "Desc": "Removes NAs from vector columns.", + "FriendlyName": "NA Drop Transform", + "ShortName": "NADrop", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "Columns to drop the NAs for", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MissingValuesRowDropper", + "Desc": "Filters out rows that contain missing values.", + "FriendlyName": "NA Filter", + "ShortName": "NAFilter", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Column", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Complement", + "Type": "Bool", + "Desc": "If true, keep only rows that contain NA values, and filter the rest.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MissingValueSubstitutor", + "Desc": "Create an output column of the same type and size of the input column, where missing values are replaced with either the default value or the mean/min/max value (for non-text columns only).", + "FriendlyName": "NA Replace Transform", + "ShortName": "NARep", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "ReplacementString", + "Type": "String", + "Desc": "Replacement value for NAs (uses default value if not given)", + "Aliases": [ + "rep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Kind", + "Type": { + "Kind": "Enum", + "Values": [ + "DefaultValue", + "Mean", + "Minimum", + "Maximum", + "SpecifiedValue", + "Mode" + ] + }, + "Desc": "The replacement method to utilize", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Slot", + "Type": "Bool", + "Desc": "Whether to impute values by slot", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:rep:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ReplacementKind", + "Type": { + "Kind": "Enum", + "Values": [ + "DefaultValue", + "Mean", + "Minimum", + "Maximum", + "SpecifiedValue", + "Mode" + ] + }, + "Desc": "The replacement method to utilize", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Def" + }, + { + "Name": "ImputeBySlot", + "Type": "Bool", + "Desc": "Whether to impute values by slot", + "Aliases": [ + "slot" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ModelCombiner", + "Desc": "Combines a sequence of TransformModels into a single model", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "TransformModel" + }, + "Desc": "Input models", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputModel", + "Type": "TransformModel", + "Desc": "Combined model" + } + ] + }, + { + "Name": "Transforms.NGramTranslator", + "Desc": "Produces a bag of counts of n-grams (sequences of consecutive values of length 1-n) in a given vector of keys. It does so by building a dictionary of n-grams and using the id in the dictionary as the index in the bag.", + "FriendlyName": "NGram Transform", + "ShortName": "NgramTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "NgramLength", + "Type": "Int", + "Desc": "Maximum n-gram length", + "Aliases": [ + "ngram" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "AllLengths", + "Type": "Bool", + "Desc": "Whether to include all n-gram lengths up to NgramLength or only NgramLength", + "Aliases": [ + "all" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "SkipLength", + "Type": "Int", + "Desc": "Maximum number of tokens to skip when constructing an n-gram", + "Aliases": [ + "skips" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxNumTerms", + "Type": { + "Kind": "Array", + "ItemType": "Int" + }, + "Desc": "Maximum number of n-grams to store in the dictionary", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Weighting", + "Type": { + "Kind": "Enum", + "Values": [ + "Tf", + "Idf", + "TfIdf" + ] + }, + "Desc": "Statistical measure used to evaluate how important a word is to a document in a corpus", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NgramLength", + "Type": "Int", + "Desc": "Maximum n-gram length", + "Aliases": [ + "ngram" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2 + }, + { + "Name": "AllLengths", + "Type": "Bool", + "Desc": "Whether to store all n-gram lengths up to ngramLength, or only ngramLength", + "Aliases": [ + "all" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "SkipLength", + "Type": "Int", + "Desc": "Maximum number of tokens to skip when constructing an n-gram", + "Aliases": [ + "skips" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MaxNumTerms", + "Type": { + "Kind": "Array", + "ItemType": "Int" + }, + "Desc": "Maximum number of n-grams to store in the dictionary", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + 10000000 + ] + }, + { + "Name": "Weighting", + "Type": { + "Kind": "Enum", + "Values": [ + "Tf", + "Idf", + "TfIdf" + ] + }, + "Desc": "The weighting criteria", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Tf" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.NoOperation", + "Desc": "Does nothing.", + "FriendlyName": "No Op", + "ShortName": "Nop", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.OptionalColumnCreator", + "Desc": "If the source column does not exist after deserialization, create a column with the right type and default values.", + "FriendlyName": "Optional Column Transform", + "ShortName": "optional", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "New column definition(s)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.PcaCalculator", + "Desc": "PCA is a dimensionality-reduction transform which computes the projection of a numeric vector onto a low-rank subspace.", + "FriendlyName": "Principal Component Analysis Transform", + "ShortName": "Pca", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "The name of the weight column", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Rank", + "Type": "Int", + "Desc": "The number of components in the PCA", + "Aliases": [ + "k" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Oversampling", + "Type": "Int", + "Desc": "Oversampling parameter for randomized PCA training", + "Aliases": [ + "over" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Center", + "Type": "Bool", + "Desc": "If enabled, data is centered to be zero mean", + "Aliases": [ + "center" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed for random number generation", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "The name of the weight column", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Rank", + "Type": "Int", + "Desc": "The number of components in the PCA", + "Aliases": [ + "k" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 20 + }, + { + "Name": "Oversampling", + "Type": "Int", + "Desc": "Oversampling parameter for randomized PCA training", + "Aliases": [ + "over" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 20 + }, + { + "Name": "Center", + "Type": "Bool", + "Desc": "If enabled, data is centered to be zero mean", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed for random number generation", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.PermutationFeatureImportance", + "Desc": "Permutation Feature Importance (PFI)", + "FriendlyName": "PFI", + "ShortName": "PFI", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The path to the model file", + "Aliases": [ + "path" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + }, + { + "Name": "UseFeatureWeightFilter", + "Type": "Bool", + "Desc": "Use feature weights to pre-filter features", + "Aliases": [ + "usefw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NumberOfExamplesToUse", + "Type": "Int", + "Desc": "Limit the number of examples to evaluate on", + "Aliases": [ + "numexamples" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "PermutationCount", + "Type": "Int", + "Desc": "The number of permutations to perform", + "Aliases": [ + "permutations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + } + ], + "Outputs": [ + { + "Name": "Metrics", + "Type": "DataView", + "Desc": "The PFI metrics" + } + ], + "InputKind": [ + "ITransformInput" + ] + }, + { + "Name": "Transforms.PredictedLabelColumnOriginalValueConverter", + "Desc": "Transforms a predicted label column to its original values, unless it is of type bool.", + "FriendlyName": "Convert Predicted Label", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictedLabelColumn", + "Type": "String", + "Desc": "The predicted label column", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RandomNumberGenerator", + "Desc": "Adds a column with a generated number sequence.", + "FriendlyName": "Generate Number Transform", + "ShortName": "Generate", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "UseCounter", + "Type": "Bool", + "Desc": "Use an auto-incremented integer starting at zero instead of a random number", + "Aliases": [ + "cnt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "The random seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:seed)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseCounter", + "Type": "Bool", + "Desc": "Use an auto-incremented integer starting at zero instead of a random number", + "Aliases": [ + "cnt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "The random seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 42 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RobustScalingNormalizer", + "Desc": "Optionally centers the data and scales based on the range of data and the quantile min and max values provided. This method is more robust to outliers.", + "FriendlyName": "Robust Scaling Normalizer", + "ShortName": "RobScal", + "Inputs": [ + { + "Name": "CenterData", + "Type": "Bool", + "Desc": "Should the data be centered around 0", + "Aliases": [ + "center" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "QuantileMin", + "Type": "UInt", + "Desc": "Minimum quantile value. Defaults to 25", + "Aliases": [ + "qmin" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 25 + }, + { + "Name": "QuantileMax", + "Type": "UInt", + "Desc": "Maximum quantile value. Defaults to 75", + "Aliases": [ + "qmax" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 75 + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RowRangeFilter", + "Desc": "Filters a dataview on a column of type Single, Double or Key (contiguous). Keeps the values that are in the specified min/max range. NaNs are always filtered out. If the input is a Key type, the min/max are considered percentages of the number of values.", + "FriendlyName": "Range Filter", + "ShortName": "RangeFilter", + "Inputs": [ + { + "Name": "Column", + "Type": "String", + "Desc": "Column", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Min", + "Type": "Float", + "Desc": "Minimum value (0 to 1 for key types)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Max", + "Type": "Float", + "Desc": "Maximum value (0 to 1 for key types)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Complement", + "Type": "Bool", + "Desc": "If true, keep the values that fall outside the range.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "IncludeMin", + "Type": "Bool", + "Desc": "If true, include in the range the values that are equal to min.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "IncludeMax", + "Type": "Bool", + "Desc": "If true, include in the range the values that are equal to max.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RowSkipAndTakeFilter", + "Desc": "Allows limiting input to a subset of rows at an optional offset. Can be used to implement data paging.", + "FriendlyName": "Skip and Take Filter", + "ShortName": "SkipTake", + "Inputs": [ + { + "Name": "Skip", + "Type": "Int", + "Desc": "Number of items to skip", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Take", + "Type": "Int", + "Desc": "Number of items to take", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RowSkipFilter", + "Desc": "Allows limiting input to a subset of rows by skipping a number of rows.", + "FriendlyName": "Skip Filter", + "ShortName": "Skip", + "Inputs": [ + { + "Name": "Count", + "Type": "Int", + "Desc": "Number of items to skip", + "Aliases": [ + "c", + "n", + "s" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RowTakeFilter", + "Desc": "Allows limiting input to a subset of rows by taking N first rows.", + "FriendlyName": "Take Filter", + "ShortName": "Take", + "Inputs": [ + { + "Name": "Count", + "Type": "Int", + "Desc": "Number of items to take", + "Aliases": [ + "c", + "n", + "t" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 9223372036854775807 + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ScoreColumnSelector", + "Desc": "Selects only the last score columns and the extra columns specified in the arguments.", + "FriendlyName": "Choose Columns By Indices", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ExtraColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Extra columns to write", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.Scorer", + "Desc": "Turn the predictor model into a transform model", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model to turn into a transform", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "ScoredData", + "Type": "DataView", + "Desc": "The scored dataset" + }, + { + "Name": "ScoringTransform", + "Type": "TransformModel", + "Desc": "The scoring transform" + } + ] + }, + { + "Name": "Transforms.Segregator", + "Desc": "Un-groups vector columns into sequences of rows, inverse of Group transform", + "FriendlyName": "Un-group Transform", + "ShortName": "Ungroup", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Columns to unroll, or 'pivot'", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + }, + { + "Name": "Mode", + "Type": { + "Kind": "Enum", + "Values": [ + "Inner", + "Outer", + "First" + ] + }, + "Desc": "Specifies how to unroll multiple pivot columns of different size.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Inner" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.SentimentAnalyzer", + "Desc": "Uses a pretrained sentiment model to score input strings", + "FriendlyName": "Sentiment Analyzing Transform", + "ShortName": "Senti", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column.", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column.", + "Aliases": [ + "dst" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.TensorFlowScorer", + "Desc": "Transforms the data using the TensorFlow model.", + "FriendlyName": "TensorFlowTransform", + "ShortName": "TFTransform", + "Inputs": [ + { + "Name": "ModelLocation", + "Type": "String", + "Desc": "TensorFlow model used by the transform. Please see https://www.tensorflow.org/mobile/prepare_models for more details.", + "Required": true, + "SortOrder": 0.0, + "IsNullable": false + }, + { + "Name": "InputColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "The names of the model inputs", + "Aliases": [ + "inputs" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "OutputColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "The name of the outputs", + "Aliases": [ + "outputs" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of samples to use for mini-batch training.", + "Required": false, + "SortOrder": 9.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "AddBatchDimensionInputs", + "Type": "Bool", + "Desc": "Add a batch dimension to the input e.g. input = [224, 224, 3] => [-1, 224, 224, 3].", + "Required": false, + "SortOrder": 16.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TreatOutputAsBatched", + "Type": "Bool", + "Desc": "If the first dimension of the output is unknown, should it be treated as batched or not. e.g. output = [-1] will be read as a vector of unknown length when this is false.", + "Required": false, + "SortOrder": 17.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.TextFeaturizer", + "Desc": "A transform that turns a collection of text documents into numerical feature vectors. The feature vectors are normalized counts of (word and/or character) n-grams in a given tokenized text.", + "FriendlyName": "Text Transform", + "ShortName": "Text", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + }, + "Desc": "New column definition (optional form: name:srcs).", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Language", + "Type": { + "Kind": "Enum", + "Values": [ + "English", + "French", + "German", + "Dutch", + "Italian", + "Spanish", + "Japanese" + ] + }, + "Desc": "Dataset language or 'AutoDetect' to detect language per row.", + "Aliases": [ + "lang" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "English" + }, + { + "Name": "StopWordsRemover", + "Type": { + "Kind": "Component", + "ComponentKind": "StopWordsRemover" + }, + "Desc": "Stopwords remover.", + "Aliases": [ + "remover" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "TextCase", + "Type": { + "Kind": "Enum", + "Values": [ + "Lower", + "Upper", + "None" + ] + }, + "Desc": "Casing text using the rules of the invariant culture.", + "Aliases": [ + "case" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Lower" + }, + { + "Name": "KeepDiacritics", + "Type": "Bool", + "Desc": "Whether to keep diacritical marks or remove them.", + "Aliases": [ + "diac" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "KeepPunctuations", + "Type": "Bool", + "Desc": "Whether to keep punctuation marks or remove them.", + "Aliases": [ + "punc" + ], + "Required": false, + "SortOrder": 7.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "KeepNumbers", + "Type": "Bool", + "Desc": "Whether to keep numbers or remove them.", + "Aliases": [ + "num" + ], + "Required": false, + "SortOrder": 8.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "OutputTokensColumnName", + "Type": "String", + "Desc": "Column containing the transformed text tokens.", + "Aliases": [ + "tokens", + "showtext", + "showTransformedText" + ], + "Required": false, + "SortOrder": 9.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Dictionary", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value, items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "ByOccurrence" + }, + { + "Name": "DropUnknowns", + "Type": "Bool", + "Desc": "Drop unknown terms instead of mapping them to NA term.", + "Aliases": [ + "dropna" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": false + } + ] + }, + "Desc": "A dictionary of allowed terms.", + "Aliases": [ + "dict" + ], + "Required": false, + "SortOrder": 10.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WordFeatureExtractor", + "Type": { + "Kind": "Component", + "ComponentKind": "NgramExtractor" + }, + "Desc": "Ngram feature extractor to use for words (WordBag/WordHashBag).", + "Aliases": [ + "wordExtractor" + ], + "Required": false, + "SortOrder": 11.0, + "IsNullable": false, + "Default": { + "Name": "NGram", + "Settings": { + "NgramLength": 2, + "MaxNumTerms": [ + 10000000 + ] + } + } + }, + { + "Name": "CharFeatureExtractor", + "Type": { + "Kind": "Component", + "ComponentKind": "NgramExtractor" + }, + "Desc": "Ngram feature extractor to use for characters (WordBag/WordHashBag).", + "Aliases": [ + "charExtractor" + ], + "Required": false, + "SortOrder": 12.0, + "IsNullable": false, + "Default": { + "Name": "NGram", + "Settings": { + "NgramLength": 3, + "AllLengths": false, + "MaxNumTerms": [ + 10000000 + ] + } + } + }, + { + "Name": "VectorNormalizer", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "L1", + "L2", + "Infinity" + ] + }, + "Desc": "Normalize vectors (rows) individually by rescaling them to unit norm.", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 13.0, + "IsNullable": false, + "Default": "L2" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.TextToKeyConverter", + "Desc": "Converts input values (words, numbers, etc.) to index in a dictionary.", + "FriendlyName": "To Key", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of terms to keep when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of keys to keep per column when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 113.0, + "IsNullable": false, + "Default": "ByOccurrence" + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 114.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.TrainTestDatasetSplitter", + "Desc": "Split the dataset into train and test sets", + "FriendlyName": "Dataset Train-Test Split", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Fraction", + "Type": "Float", + "Desc": "Fraction of training data", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0.8 + }, + { + "Name": "StratificationColumn", + "Type": "String", + "Desc": "Stratification column", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "TrainData", + "Type": "DataView", + "Desc": "Training data" + }, + { + "Name": "TestData", + "Type": "DataView", + "Desc": "Testing data" + } + ] + }, + { + "Name": "Transforms.TreeLeafFeaturizer", + "Desc": "Trains a tree ensemble, or loads it from a file, then maps a numeric feature vector to three outputs: 1. A vector containing the individual tree outputs of the tree ensemble. 2. A vector indicating the leaves that the feature vector falls on in the tree ensemble. 3. A vector indicating the paths that the feature vector falls on in the tree ensemble. If a both a model file and a trainer are specified - will use the model file. If neither are specified, will train a default FastTree model. This can handle key labels by training a regression model towards their optionally permuted indices.", + "FriendlyName": "Tree Ensemble Featurization Transform", + "ShortName": "TreeFeat", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Trainer to use", + "Required": true, + "SortOrder": 10.0, + "IsNullable": false + }, + { + "Name": "Suffix", + "Type": "String", + "Desc": "Output column: The suffix to append to the default column names", + "Aliases": [ + "ex" + ], + "Required": false, + "SortOrder": 101.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "LabelPermutationSeed", + "Type": "Int", + "Desc": "If specified, determines the permutation seed for applying this featurizer to a multiclass problem.", + "Aliases": [ + "lps" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "IFeaturizerInput", + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.TwoHeterogeneousModelCombiner", + "Desc": "Combines a TransformModel and a PredictorModel into a single PredictorModel.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "TransformModel", + "Type": "TransformModel", + "Desc": "Transform model", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Predictor model", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Predictor model" + } + ] + }, + { + "Name": "Transforms.VectorToImage", + "Desc": "Converts vector array into image type.", + "FriendlyName": "Vector To Image Transform", + "ShortName": "VectorToImageConverter", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "ContainsAlpha", + "Type": "Bool", + "Desc": "Whether to use alpha channel", + "Aliases": [ + "alpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ContainsRed", + "Type": "Bool", + "Desc": "Whether to use red channel", + "Aliases": [ + "red" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ContainsGreen", + "Type": "Bool", + "Desc": "Whether to use green channel", + "Aliases": [ + "green" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ContainsBlue", + "Type": "Bool", + "Desc": "Whether to use blue channel", + "Aliases": [ + "blue" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Order", + "Type": { + "Kind": "Enum", + "Values": [ + "ARGB", + "ARBG", + "ABRG", + "ABGR", + "AGRB", + "AGBR" + ] + }, + "Desc": "Order of channels", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Interleave", + "Type": "Bool", + "Desc": "Whether to separate each channel or interleave in specified order", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ImageWidth", + "Type": "Int", + "Desc": "Width of the image", + "Aliases": [ + "width" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ImageHeight", + "Type": "Int", + "Desc": "Height of the image", + "Aliases": [ + "height" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "Offset (pre-scale)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale factor", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DefaultAlpha", + "Type": "Int", + "Desc": "Default value for alpha channel. Will be used if ContainsAlpha set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DefaultRed", + "Type": "Int", + "Desc": "Default value for red channel. Will be used if ContainsRed set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DefaultGreen", + "Type": "Int", + "Desc": "Default value for green channel. Will be used if ContainsGreen set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DefaultBlue", + "Type": "Int", + "Desc": "Default value for blue channel. Will be used if ContainsGreen set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ContainsAlpha", + "Type": "Bool", + "Desc": "Whether to use alpha channel", + "Aliases": [ + "alpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "ContainsRed", + "Type": "Bool", + "Desc": "Whether to use red channel", + "Aliases": [ + "red" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "ContainsGreen", + "Type": "Bool", + "Desc": "Whether to use green channel", + "Aliases": [ + "green" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "ContainsBlue", + "Type": "Bool", + "Desc": "Whether to use blue channel", + "Aliases": [ + "blue" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Order", + "Type": { + "Kind": "Enum", + "Values": [ + "ARGB", + "ARBG", + "ABRG", + "ABGR", + "AGRB", + "AGBR" + ] + }, + "Desc": "Order of colors.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "ARGB" + }, + { + "Name": "Interleave", + "Type": "Bool", + "Desc": "Whether to separate each channel or interleave in specified order", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "ImageWidth", + "Type": "Int", + "Desc": "Width of the image", + "Aliases": [ + "width" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "ImageHeight", + "Type": "Int", + "Desc": "Height of the image", + "Aliases": [ + "height" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "Offset (pre-scale)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale factor", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "DefaultAlpha", + "Type": "Int", + "Desc": "Default value for alpha channel. Will be used if ContainsAlpha set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "DefaultRed", + "Type": "Int", + "Desc": "Default value for red channel. Will be used if ContainsRed set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "DefaultGreen", + "Type": "Int", + "Desc": "Default value for green channel. Will be used if ContainsGreen set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "DefaultBlue", + "Type": "Int", + "Desc": "Default value for blue channel. Will be used if ContainsBlue set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.WordEmbeddings", + "Desc": "Word Embeddings transform is a text featurizer which converts vectors of text tokens into sentence vectors using a pre-trained model", + "FriendlyName": "Word Embeddings Transform", + "ShortName": "WordEmbeddings", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 0.0, + "IsNullable": false + }, + { + "Name": "ModelKind", + "Type": { + "Kind": "Enum", + "Values": [ + "GloVe50D", + "GloVe100D", + "GloVe200D", + "GloVe300D", + "GloVeTwitter25D", + "GloVeTwitter50D", + "GloVeTwitter100D", + "GloVeTwitter200D", + "FastTextWikipedia300D", + "SentimentSpecificWordEmbedding" + ] + }, + "Desc": "Pre-trained model used to create the vocabulary", + "Aliases": [ + "model" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": "SentimentSpecificWordEmbedding" + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "CustomLookupTable", + "Type": "String", + "Desc": "Filename for custom word embedding model", + "Aliases": [ + "dataFile" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.WordTokenizer", + "Desc": "The input to this transform is text, and the output is a vector of text containing the words (tokens) in the original text. The separator is space, but can be specified as any other character (or multiple characters) if needed.", + "FriendlyName": "Tokenize Text Transform", + "ShortName": "TokenizeTextTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "TermSeparators", + "Type": "String", + "Desc": "Comma separated set of term separator(s). Commonly: 'space', 'comma', 'semicolon' or other single character.", + "Aliases": [ + "sep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "CharArrayTermSeparators", + "Type": { + "Kind": "Array", + "ItemType": "Char" + }, + "Desc": "Array of single character term separator(s). By default uses space character separator.", + "Aliases": [ + "sep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + } + ], + "Components": [ + { + "Kind": "BoosterParameterFunction", + "Components": [ + { + "Name": "dart", + "Desc": "Dropouts meet Multiple Additive Regresion Trees. See https://arxiv.org/abs/1505.01866", + "FriendlyName": "Tree Dropout Tree Booster", + "Settings": [ + { + "Name": "TreeDropFraction", + "Type": "Float", + "Desc": "The drop ratio for trees. Range:(0,1).", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "MaximumNumberOfDroppedTreesPerRound", + "Type": "Int", + "Desc": "Maximum number of dropped trees in a boosting round.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1, + "Range": { + "Inf": 0, + "Max": 2147483647 + } + }, + { + "Name": "SkipDropFraction", + "Type": "Float", + "Desc": "Probability for not dropping in a boosting round.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.5, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "XgboostDartMode", + "Type": "Bool", + "Desc": "True will enable xgboost dart mode.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UniformDrop", + "Type": "Bool", + "Desc": "True will enable uniform drop.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MinimumSplitGain", + "Type": "Float", + "Desc": "Minimum loss reduction required to make a further partition on a leaf node of the tree. the larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "MaximumTreeDepth", + "Type": "Int", + "Desc": "Maximum depth of a tree. 0 means no limit. However, tree still grows by best-first.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "MinimumChildWeight", + "Type": "Float", + "Desc": "Minimum sum of instance weight(hessian) needed in a child. If the tree partition step results in a leaf node with the sum of instance weight less than min_child_weight, then the building process will give up further partitioning. In linear regression mode, this simply corresponds to minimum number of instances needed to be in each node. The larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "SubsampleFrequency", + "Type": "Int", + "Desc": "Subsample frequency for bagging. 0 means no subsample. Specifies the frequency at which the bagging occurs, where if this is set to N, the subsampling will happen at every N iterations.This must be set with Subsample as this specifies the amount to subsample.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "SubsampleFraction", + "Type": "Float", + "Desc": "Subsample ratio of the training instance. Setting it to 0.5 means that LightGBM randomly collected half of the data instances to grow trees and this will prevent overfitting. Range: (0,1].", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "Subsample ratio of columns when constructing each tree. Range: (0,1].", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization term on weights, increasing this value will make model more conservative.", + "Aliases": [ + "l2" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization term on weights, increase this value will make model more conservative.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + } + ] + }, + { + "Name": "gbdt", + "Desc": "Traditional Gradient Boosting Decision Tree.", + "FriendlyName": "Tree Booster", + "Settings": [ + { + "Name": "MinimumSplitGain", + "Type": "Float", + "Desc": "Minimum loss reduction required to make a further partition on a leaf node of the tree. the larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "MaximumTreeDepth", + "Type": "Int", + "Desc": "Maximum depth of a tree. 0 means no limit. However, tree still grows by best-first.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "MinimumChildWeight", + "Type": "Float", + "Desc": "Minimum sum of instance weight(hessian) needed in a child. If the tree partition step results in a leaf node with the sum of instance weight less than min_child_weight, then the building process will give up further partitioning. In linear regression mode, this simply corresponds to minimum number of instances needed to be in each node. The larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "SubsampleFrequency", + "Type": "Int", + "Desc": "Subsample frequency for bagging. 0 means no subsample. Specifies the frequency at which the bagging occurs, where if this is set to N, the subsampling will happen at every N iterations.This must be set with Subsample as this specifies the amount to subsample.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "SubsampleFraction", + "Type": "Float", + "Desc": "Subsample ratio of the training instance. Setting it to 0.5 means that LightGBM randomly collected half of the data instances to grow trees and this will prevent overfitting. Range: (0,1].", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "Subsample ratio of columns when constructing each tree. Range: (0,1].", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization term on weights, increasing this value will make model more conservative.", + "Aliases": [ + "l2" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization term on weights, increase this value will make model more conservative.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + } + ] + }, + { + "Name": "goss", + "Desc": "Gradient-based One-Side Sampling.", + "FriendlyName": "Gradient-based One-Size Sampling", + "Settings": [ + { + "Name": "TopRate", + "Type": "Float", + "Desc": "Retain ratio for large gradient instances.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.2, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "OtherRate", + "Type": "Float", + "Desc": "Retain ratio for small gradient instances.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "MinimumSplitGain", + "Type": "Float", + "Desc": "Minimum loss reduction required to make a further partition on a leaf node of the tree. the larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "MaximumTreeDepth", + "Type": "Int", + "Desc": "Maximum depth of a tree. 0 means no limit. However, tree still grows by best-first.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "MinimumChildWeight", + "Type": "Float", + "Desc": "Minimum sum of instance weight(hessian) needed in a child. If the tree partition step results in a leaf node with the sum of instance weight less than min_child_weight, then the building process will give up further partitioning. In linear regression mode, this simply corresponds to minimum number of instances needed to be in each node. The larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "SubsampleFrequency", + "Type": "Int", + "Desc": "Subsample frequency for bagging. 0 means no subsample. Specifies the frequency at which the bagging occurs, where if this is set to N, the subsampling will happen at every N iterations.This must be set with Subsample as this specifies the amount to subsample.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "SubsampleFraction", + "Type": "Float", + "Desc": "Subsample ratio of the training instance. Setting it to 0.5 means that LightGBM randomly collected half of the data instances to grow trees and this will prevent overfitting. Range: (0,1].", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "Subsample ratio of columns when constructing each tree. Range: (0,1].", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization term on weights, increasing this value will make model more conservative.", + "Aliases": [ + "l2" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization term on weights, increase this value will make model more conservative.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + } + ] + } + ] + }, + { + "Kind": "CalibratorTrainer", + "Components": [ + { + "Name": "FixedPlattCalibrator", + "Desc": null, + "FriendlyName": "Fixed Platt Calibrator", + "Aliases": [ + "FixedPlatt", + "FixedSigmoid" + ], + "Settings": [ + { + "Name": "Slope", + "Type": "Float", + "Desc": "The slope parameter of f(x) = 1 / (1 + exp(slope * x + offset)", + "Aliases": [ + "a" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1.0 + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "The offset parameter of f(x) = 1 / (1 + exp(slope * x + offset)", + "Aliases": [ + "b" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + } + ] + }, + { + "Name": "NaiveCalibrator", + "Desc": null, + "FriendlyName": "Naive Calibrator", + "Aliases": [ + "Naive" + ], + "Settings": [] + }, + { + "Name": "PavCalibrator", + "Desc": null, + "FriendlyName": "PAV Calibrator", + "Aliases": [ + "Pav" + ], + "Settings": [] + }, + { + "Name": "PlattCalibrator", + "Desc": "Platt calibration.", + "FriendlyName": "Platt Calibrator", + "Aliases": [ + "Platt", + "Sigmoid" + ], + "Settings": [] + } + ] + }, + { + "Kind": "ClassificationLossFunction", + "Components": [ + { + "Name": "ExpLoss", + "Desc": "Exponential loss.", + "FriendlyName": "Exponential Loss", + "Settings": [ + { + "Name": "Beta", + "Type": "Float", + "Desc": "Beta (dilation)", + "Aliases": [ + "beta" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ] + }, + { + "Name": "HingeLoss", + "Desc": "Hinge loss.", + "FriendlyName": "Hinge loss", + "Aliases": [ + "Hinge" + ], + "Settings": [ + { + "Name": "Margin", + "Type": "Float", + "Desc": "Margin value", + "Aliases": [ + "marg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ] + }, + { + "Name": "LogLoss", + "Desc": "Log loss.", + "FriendlyName": "Log loss", + "Aliases": [ + "Logistic", + "CrossEntropy" + ], + "Settings": [] + }, + { + "Name": "SmoothedHingeLoss", + "Desc": "Smoothed Hinge loss.", + "FriendlyName": "Smoothed Hinge Loss", + "Aliases": [ + "SmoothedHinge" + ], + "Settings": [ + { + "Name": "SmoothingConst", + "Type": "Float", + "Desc": "Smoothing constant", + "Aliases": [ + "smooth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ] + } + ] + }, + { + "Kind": "CountTableBuilder", + "Components": [ + { + "Name": "CMSketch", + "Desc": "Create the count table using the count-min sketch structure, which has a smaller memory footprint, at the expense of some overcounting due to collisions.", + "FriendlyName": "Count Min Table Builder", + "Aliases": [ + "CMTable" + ], + "Settings": [ + { + "Name": "Depth", + "Type": "Int", + "Desc": "Count-Min Sketch table depth", + "Aliases": [ + "d" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 4 + }, + { + "Name": "Width", + "Type": "Int", + "Desc": "Count-Min Sketch width", + "Aliases": [ + "w" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 8388608 + } + ] + }, + { + "Name": "Dict", + "Desc": "Build a dictionary containing the exact count of each categorical feature value.", + "FriendlyName": "Dictionary Based Count Table Builder", + "Aliases": [ + "Dictionary" + ], + "Settings": [ + { + "Name": "GarbageThreshold", + "Type": "Float", + "Desc": "Garbage threshold (counts below or equal to the threshold are assigned to the garbage bin)", + "Aliases": [ + "gb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + } + ] + } + ] + }, + { + "Kind": "EarlyStoppingCriterion", + "Components": [ + { + "Name": "GL", + "Desc": "Stop in case of loss of generality.", + "FriendlyName": "Loss of Generality (GL)", + "Settings": [ + { + "Name": "Threshold", + "Type": "Float", + "Desc": "Threshold in range [0,1].", + "Aliases": [ + "th" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Max": 1.0, + "Min": 0.0 + } + } + ] + }, + { + "Name": "LP", + "Desc": "Stops in case of low progress.", + "FriendlyName": "Low Progress (LP)", + "Settings": [ + { + "Name": "Threshold", + "Type": "Float", + "Desc": "Threshold in range [0,1].", + "Aliases": [ + "th" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Max": 1.0, + "Min": 0.0 + } + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The window size.", + "Aliases": [ + "w" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5, + "Range": { + "Inf": 0 + } + } + ] + }, + { + "Name": "PQ", + "Desc": "Stops in case of generality to progress ration exceeds threshold.", + "FriendlyName": "Generality to Progress Ratio (PQ)", + "Settings": [ + { + "Name": "Threshold", + "Type": "Float", + "Desc": "Threshold in range [0,1].", + "Aliases": [ + "th" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Max": 1.0, + "Min": 0.0 + } + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The window size.", + "Aliases": [ + "w" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5, + "Range": { + "Inf": 0 + } + } + ] + }, + { + "Name": "TR", + "Desc": "Stop if validation score exceeds threshold value.", + "FriendlyName": "Tolerant (TR)", + "Settings": [ + { + "Name": "Threshold", + "Type": "Float", + "Desc": "Tolerance threshold. (Non negative value)", + "Aliases": [ + "th" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Min": 0.0 + } + } + ] + }, + { + "Name": "UP", + "Desc": "Stops in case of consecutive loss in generality.", + "FriendlyName": "Consecutive Loss in Generality (UP)", + "Settings": [ + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The window size.", + "Aliases": [ + "w" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5, + "Range": { + "Inf": 0 + } + } + ] + } + ] + }, + { + "Kind": "EnsembleBinaryDiversityMeasure", + "Components": [ + { + "Name": "DisagreementDiversityMeasure", + "Desc": null, + "FriendlyName": "Disagreement Diversity Measure", + "Settings": [] + } + ] + }, + { + "Kind": "EnsembleBinaryOutputCombiner", + "Components": [ + { + "Name": "Average", + "Desc": null, + "FriendlyName": "Average", + "Settings": [] + }, + { + "Name": "Median", + "Desc": null, + "FriendlyName": "Median", + "Settings": [] + }, + { + "Name": "Stacking", + "Desc": null, + "FriendlyName": "Stacking", + "Settings": [ + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + }, + { + "Name": "Voting", + "Desc": null, + "FriendlyName": "Voting", + "Settings": [] + }, + { + "Name": "WeightedAverage", + "Desc": null, + "FriendlyName": "Weighted Average", + "Settings": [ + { + "Name": "WeightageName", + "Type": { + "Kind": "Enum", + "Values": [ + "Accuracy", + "Auc", + "PosPrecision", + "PosRecall", + "NegPrecision", + "NegRecall" + ] + }, + "Desc": "The metric type to be used to find the weights for each model", + "Aliases": [ + "wn" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": "Auc" + } + ] + } + ] + }, + { + "Kind": "EnsembleBinarySubModelSelector", + "Components": [ + { + "Name": "AllSelector", + "Desc": null, + "FriendlyName": "All Selector", + "Settings": [] + }, + { + "Name": "BestDiverseSelector", + "Desc": null, + "FriendlyName": "Best Diverse Selector", + "Settings": [ + { + "Name": "DiversityMetricType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleBinaryDiversityMeasure" + }, + "Desc": "The metric type to be used to find the diversity among base learners", + "Aliases": [ + "dm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "DisagreementDiversityMeasure" + } + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + }, + { + "Name": "BestPerformanceSelector", + "Desc": null, + "FriendlyName": "Best Performance Selector", + "Settings": [ + { + "Name": "MetricName", + "Type": { + "Kind": "Enum", + "Values": [ + "Accuracy", + "PosPrecName", + "PosRecallName", + "NegPrecName", + "NegRecallName", + "Auc", + "LogLoss", + "LogLossReduction", + "F1", + "AuPrc" + ] + }, + "Desc": "The metric type to be used to find the best performance", + "Aliases": [ + "mn" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": "Auc" + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + } + ] + }, + { + "Kind": "EnsembleFeatureSelector", + "Components": [ + { + "Name": "AllFeatureSelector", + "Desc": null, + "FriendlyName": "All Feature Selector", + "Settings": [] + }, + { + "Name": "RandomFeatureSelector", + "Desc": null, + "FriendlyName": "Random Feature Selector", + "Settings": [ + { + "Name": "FeaturesSelectionProportion", + "Type": "Float", + "Desc": "The proportion of features to be selected. The range is 0.0-1.0", + "Aliases": [ + "fp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.8 + } + ] + } + ] + }, + { + "Kind": "EnsembleMulticlassDiversityMeasure", + "Components": [ + { + "Name": "MultiDisagreementDiversityMeasure", + "Desc": null, + "FriendlyName": "Disagreement Diversity Measure", + "Settings": [] + } + ] + }, + { + "Kind": "EnsembleMulticlassOutputCombiner", + "Components": [ + { + "Name": "MultiAverage", + "Desc": null, + "FriendlyName": "Average", + "Settings": [ + { + "Name": "Normalize", + "Type": "Bool", + "Desc": "Whether to normalize the output of base models before combining them", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": true + } + ] + }, + { + "Name": "MultiMedian", + "Desc": null, + "FriendlyName": "Median", + "Settings": [ + { + "Name": "Normalize", + "Type": "Bool", + "Desc": "Whether to normalize the output of base models before combining them", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": true + } + ] + }, + { + "Name": "MultiStacking", + "Desc": null, + "FriendlyName": "Stacking", + "Settings": [ + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + }, + { + "Name": "MultiVoting", + "Desc": null, + "FriendlyName": "Voting", + "Settings": [] + }, + { + "Name": "MultiWeightedAverage", + "Desc": null, + "FriendlyName": "Multi Weighted Average", + "Settings": [ + { + "Name": "WeightageName", + "Type": { + "Kind": "Enum", + "Values": [ + "AccuracyMicroAvg", + "AccuracyMacroAvg" + ] + }, + "Desc": "The metric type to be used to find the weights for each model", + "Aliases": [ + "wn" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": "AccuracyMicroAvg" + }, + { + "Name": "Normalize", + "Type": "Bool", + "Desc": "Whether to normalize the output of base models before combining them", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": true + } + ] + } + ] + }, + { + "Kind": "EnsembleMulticlassSubModelSelector", + "Components": [ + { + "Name": "AllSelectorMultiClass", + "Desc": null, + "FriendlyName": "All Selector", + "Settings": [] + }, + { + "Name": "BestDiverseSelectorMultiClass", + "Desc": null, + "FriendlyName": "Best Diverse Selector", + "Settings": [ + { + "Name": "DiversityMetricType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleMulticlassDiversityMeasure" + }, + "Desc": "The metric type to be used to find the diversity among base learners", + "Aliases": [ + "dm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "MultiDisagreementDiversityMeasure" + } + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + }, + { + "Name": "BestPerformanceSelectorMultiClass", + "Desc": null, + "FriendlyName": "Best Performance Selector", + "Settings": [ + { + "Name": "MetricName", + "Type": { + "Kind": "Enum", + "Values": [ + "AccuracyMicro", + "AccuracyMacro", + "LogLoss", + "LogLossReduction" + ] + }, + "Desc": "The metric type to be used to find the best performance", + "Aliases": [ + "mn" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": "AccuracyMicro" + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + } + ] + }, + { + "Kind": "EnsembleRegressionDiversityMeasure", + "Components": [ + { + "Name": "RegressionDisagreementDiversityMeasure", + "Desc": null, + "FriendlyName": "Disagreement Diversity Measure", + "Settings": [] + } + ] + }, + { + "Kind": "EnsembleRegressionOutputCombiner", + "Components": [ + { + "Name": "Average", + "Desc": null, + "FriendlyName": "Average", + "Settings": [] + }, + { + "Name": "Median", + "Desc": null, + "FriendlyName": "Median", + "Settings": [] + }, + { + "Name": "RegressionStacking", + "Desc": null, + "FriendlyName": "Stacking", + "Settings": [ + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + } + ] + }, + { + "Kind": "EnsembleRegressionSubModelSelector", + "Components": [ + { + "Name": "AllSelector", + "Desc": null, + "FriendlyName": "All Selector", + "Settings": [] + }, + { + "Name": "BestDiverseSelectorRegression", + "Desc": null, + "FriendlyName": "Best Diverse Selector", + "Settings": [ + { + "Name": "DiversityMetricType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleRegressionDiversityMeasure" + }, + "Desc": "The metric type to be used to find the diversity among base learners", + "Aliases": [ + "dm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "RegressionDisagreementDiversityMeasure" + } + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + }, + { + "Name": "BestPerformanceRegressionSelector", + "Desc": null, + "FriendlyName": "Best Performance Selector", + "Settings": [ + { + "Name": "MetricName", + "Type": { + "Kind": "Enum", + "Values": [ + "L1", + "L2", + "Rms", + "Loss", + "RSquared" + ] + }, + "Desc": "The metric type to be used to find the best performance", + "Aliases": [ + "mn" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": "L1" + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + } + ] + }, + { + "Kind": "EnsembleSubsetSelector", + "Components": [ + { + "Name": "AllInstanceSelector", + "Desc": null, + "FriendlyName": "All Instance Selector", + "Settings": [ + { + "Name": "FeatureSelector", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleFeatureSelector" + }, + "Desc": "The Feature selector", + "Aliases": [ + "fs" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": { + "Name": "AllFeatureSelector" + } + } + ] + }, + { + "Name": "BootstrapSelector", + "Desc": null, + "FriendlyName": "Bootstrap Selector", + "Settings": [ + { + "Name": "FeatureSelector", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleFeatureSelector" + }, + "Desc": "The Feature selector", + "Aliases": [ + "fs" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": { + "Name": "AllFeatureSelector" + } + } + ] + }, + { + "Name": "RandomPartitionSelector", + "Desc": null, + "FriendlyName": "Random Partition Selector", + "Settings": [ + { + "Name": "FeatureSelector", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleFeatureSelector" + }, + "Desc": "The Feature selector", + "Aliases": [ + "fs" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": { + "Name": "AllFeatureSelector" + } + } + ] + } + ] + }, + { + "Kind": "FastTreeTrainer", + "Components": [ + { + "Name": "FastTreeBinaryClassification", + "Desc": "Uses a logit-boost boosted tree learner to perform binary classification.", + "FriendlyName": "FastTree (Boosted Trees) Classification", + "Settings": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "UnbalancedSets", + "Type": "Bool", + "Desc": "Option for using derivatives optimized for unbalanced sets", + "Aliases": [ + "us" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ] + }, + { + "Name": "FastTreeRanking", + "Desc": "Trains gradient boosted decision trees to the LambdaRank quasi-gradient.", + "FriendlyName": "FastTree (Boosted Trees) Ranking", + "Settings": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "CustomGains", + "Type": { + "Kind": "Array", + "ItemType": "Float" + }, + "Desc": "Comma-separated list of gains associated to each relevance label.", + "Aliases": [ + "gains" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + 0.0, + 3.0, + 7.0, + 15.0, + 31.0 + ] + }, + { + "Name": "UseDcg", + "Type": "Bool", + "Desc": "Train DCG instead of NDCG", + "Aliases": [ + "dcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "SortingAlgorithm", + "Type": "String", + "Desc": "The sorting algorithm to use for DCG and LambdaMart calculations [DescendingStablePessimistic/DescendingStable/DescendingReverse/DescendingDotNet]", + "Aliases": [ + "sort" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "DescendingStablePessimistic" + }, + { + "Name": "NdcgTruncationLevel", + "Type": "Int", + "Desc": "max-NDCG truncation to use in the LambdaMART algorithm", + "Aliases": [ + "n" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "ShiftedNdcg", + "Type": "Bool", + "Desc": "Use shifted NDCG", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "CostFunctionParam", + "Type": "Char", + "Desc": "Cost function parameter (w/c)", + "Aliases": [ + "cf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "w" + }, + { + "Name": "DistanceWeight2", + "Type": "Bool", + "Desc": "Distance weight 2 adjustment to cost", + "Aliases": [ + "dw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NormalizeQueryLambdas", + "Type": "Bool", + "Desc": "Normalize query lambdas", + "Aliases": [ + "nql" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ] + }, + { + "Name": "FastTreeRegression", + "Desc": "Trains gradient boosted decision trees to fit target values using least-squares.", + "FriendlyName": "FastTree (Boosted Trees) Regression", + "Settings": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ] + }, + { + "Name": "FastTreeTweedieRegression", + "Desc": "Trains gradient boosted decision trees to fit target values using a Tweedie loss function. This learner is a generalization of Poisson, compound Poisson, and gamma regression.", + "FriendlyName": "FastTree (Boosted Trees) Tweedie Regression", + "Settings": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Index", + "Type": "Float", + "Desc": "Index parameter for the Tweedie distribution, in the range [1, 2]. 1 is Poisson loss, 2 is gamma loss, and intermediate values are compound Poisson loss.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.5 + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ] + } + ] + }, + { + "Kind": "NgramExtractor", + "Components": [ + { + "Name": "NGram", + "Desc": "Extracts NGrams from text and convert them to vector using dictionary.", + "FriendlyName": "NGram Extractor Transform", + "Aliases": [ + "NGramExtractorTransform", + "NGramExtractor" + ], + "Settings": [ + { + "Name": "NgramLength", + "Type": "Int", + "Desc": "Ngram length", + "Aliases": [ + "ngram" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "SkipLength", + "Type": "Int", + "Desc": "Maximum number of tokens to skip when constructing an n-gram", + "Aliases": [ + "skips" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "AllLengths", + "Type": "Bool", + "Desc": "Whether to include all n-gram lengths up to NgramLength or only NgramLength", + "Aliases": [ + "all" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxNumTerms", + "Type": { + "Kind": "Array", + "ItemType": "Int" + }, + "Desc": "Maximum number of n-grams to store in the dictionary", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + 10000000 + ] + }, + { + "Name": "Weighting", + "Type": { + "Kind": "Enum", + "Values": [ + "Tf", + "Idf", + "TfIdf" + ] + }, + "Desc": "The weighting criteria", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Tf" + }, + { + "Name": "TermSeparator", + "Type": "Char", + "Desc": "Separator used to separate terms/frequency pairs.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "\u0000" + }, + { + "Name": "FreqSeparator", + "Type": "Char", + "Desc": "Separator used to separate terms from their frequency.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "\u0000" + } + ] + }, + { + "Name": "NGramHash", + "Desc": "Extracts NGrams from text and convert them to vector using hashing trick.", + "FriendlyName": "NGram Hash Extractor Transform", + "Aliases": [ + "NGramHashExtractorTransform", + "NGramHashExtractor" + ], + "Settings": [ + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "Number of bits to hash into. Must be between 1 and 30, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 16 + }, + { + "Name": "NgramLength", + "Type": "Int", + "Desc": "Ngram length", + "Aliases": [ + "ngram" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "SkipLength", + "Type": "Int", + "Desc": "Maximum number of tokens to skip when constructing an n-gram", + "Aliases": [ + "skips" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "AllLengths", + "Type": "Bool", + "Desc": "Whether to include all n-gram lengths up to ngramLength or only ngramLength", + "Aliases": [ + "all" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + }, + { + "Name": "Ordered", + "Type": "Bool", + "Desc": "Whether the position of each source column should be included in the hash (when there are multiple source columns).", + "Aliases": [ + "ord" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaximumNumberOfInverts", + "Type": "Int", + "Desc": "Limit the number of keys used to generate the slot name to this many. 0 means no invert hashing, -1 means no limit.", + "Aliases": [ + "ih" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ] + } + ] + }, + { + "Kind": "ParallelLightGBM", + "Components": [ + { + "Name": "Single", + "Desc": "Single node machine learning process.", + "FriendlyName": "Single", + "Settings": [] + } + ] + }, + { + "Kind": "ParallelTraining", + "Components": [ + { + "Name": "Single", + "Desc": "Single node machine learning process.", + "FriendlyName": "Single", + "Settings": [] + } + ] + }, + { + "Kind": "PartitionedPathParser", + "Components": [ + { + "Name": "ParquetPathParser", + "Desc": "Extract name/value pairs from Parquet formatted directory names. Example path: Year=2018/Month=12/data1.parquet", + "FriendlyName": "Parquet Partitioned Path Parser", + "Aliases": [ + "ParqPP" + ], + "Settings": [] + }, + { + "Name": "SimplePathParser", + "Desc": "A simple parser that extracts directory names as column values. Column names are defined as arguments.", + "FriendlyName": "Simple Partitioned Path Parser", + "Aliases": [ + "SmplPP" + ], + "Settings": [ + { + "Name": "Columns", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the column.", + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + }, + { + "Name": "Type", + "Type": { + "Kind": "Enum", + "Values": [ + "I1", + "U1", + "I2", + "U2", + "I4", + "U4", + "I8", + "U8", + "R4", + "Num", + "R8", + "TX", + "Text", + "TXT", + "BL", + "Bool", + "TimeSpan", + "TS", + "DT", + "DateTime", + "DZ", + "DateTimeZone", + "UG", + "U16" + ] + }, + "Desc": "Data type of the column.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Source", + "Type": "Int", + "Desc": "Index of the directory representing this column.", + "Required": true, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ] + } + }, + "Desc": "Column definitions used to override the Partitioned Path Parser. Expected with the format name:type:numeric-source, for example, col=MyFeature:R4:1", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Type", + "Type": { + "Kind": "Enum", + "Values": [ + "I1", + "U1", + "I2", + "U2", + "I4", + "U4", + "I8", + "U8", + "R4", + "Num", + "R8", + "TX", + "Text", + "TXT", + "BL", + "Bool", + "TimeSpan", + "TS", + "DT", + "DateTime", + "DZ", + "DateTimeZone", + "UG", + "U16" + ] + }, + "Desc": "Data type of each column.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "TX" + } + ] + } + ] + }, + { + "Kind": "RegressionLossFunction", + "Components": [ + { + "Name": "PoissonLoss", + "Desc": "Poisson loss.", + "FriendlyName": "Poisson Loss", + "Settings": [] + }, + { + "Name": "SquaredLoss", + "Desc": "Squared loss.", + "FriendlyName": "Squared Loss", + "Aliases": [ + "L2" + ], + "Settings": [] + }, + { + "Name": "TweedieLoss", + "Desc": "Tweedie loss.", + "FriendlyName": "Tweedie Loss", + "Aliases": [ + "tweedie" + ], + "Settings": [ + { + "Name": "Index", + "Type": "Float", + "Desc": "Index parameter for the Tweedie distribution, in the range [1, 2]. 1 is Poisson loss, 2 is gamma loss, and intermediate values are compound Poisson loss.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.5 + } + ] + } + ] + }, + { + "Kind": "SDCAClassificationLossFunction", + "Components": [ + { + "Name": "HingeLoss", + "Desc": "Hinge loss.", + "FriendlyName": "Hinge loss", + "Aliases": [ + "Hinge" + ], + "Settings": [ + { + "Name": "Margin", + "Type": "Float", + "Desc": "Margin value", + "Aliases": [ + "marg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ] + }, + { + "Name": "LogLoss", + "Desc": "Log loss.", + "FriendlyName": "Log loss", + "Aliases": [ + "Logistic", + "CrossEntropy" + ], + "Settings": [] + }, + { + "Name": "SmoothedHingeLoss", + "Desc": "Smoothed Hinge loss.", + "FriendlyName": "Smoothed Hinge Loss", + "Aliases": [ + "SmoothedHinge" + ], + "Settings": [ + { + "Name": "SmoothingConst", + "Type": "Float", + "Desc": "Smoothing constant", + "Aliases": [ + "smooth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ] + } + ] + }, + { + "Kind": "SDCARegressionLossFunction", + "Components": [ + { + "Name": "SquaredLoss", + "Desc": "Squared loss.", + "FriendlyName": "Squared Loss", + "Aliases": [ + "L2" + ], + "Settings": [] + } + ] + }, + { + "Kind": "StopWordsRemover", + "Components": [ + { + "Name": "Custom", + "Desc": "Remover with list of stopwords specified by the user.", + "FriendlyName": "Custom Stopwords Remover", + "Aliases": [ + "CustomStopWordsRemover", + "CustomStopWords" + ], + "Settings": [ + { + "Name": "Stopword", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of stopwords", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + }, + { + "Name": "Predefined", + "Desc": "Remover with predefined list of stop words.", + "FriendlyName": "Predefined Stopwords List Remover", + "Aliases": [ + "PredefinedStopWordsRemover", + "PredefinedStopWords" + ], + "Settings": [] + } + ] + } + ], + "EntryPointKinds": [ + { + "Kind": "IAnomalyDetectionOutput", + "Settings": [] + }, + { + "Kind": "IBinaryClassificationOutput", + "Settings": [] + }, + { + "Kind": "ICalibratorInput", + "Settings": [ + { + "Name": "UncalibratedPredictorModel", + "Type": "PredictorModel" + }, + { + "Name": "MaxRows", + "Type": "Int" + }, + { + "Name": "Data", + "Type": "DataView" + } + ] + }, + { + "Kind": "ICalibratorOutput", + "Settings": [] + }, + { + "Kind": "IClassificationEvaluatorOutput", + "Settings": [ + { + "Name": "ConfusionMatrix", + "Type": "DataView" + }, + { + "Name": "Warnings", + "Type": "DataView" + }, + { + "Name": "OverallMetrics", + "Type": "DataView" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView" + } + ] + }, + { + "Kind": "IClusteringOutput", + "Settings": [] + }, + { + "Kind": "IEvaluatorInput", + "Settings": [ + { + "Name": "Data", + "Type": "DataView" + }, + { + "Name": "NameColumn", + "Type": "String" + } + ] + }, + { + "Kind": "IEvaluatorOutput", + "Settings": [ + { + "Name": "Warnings", + "Type": "DataView" + }, + { + "Name": "OverallMetrics", + "Type": "DataView" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView" + } + ] + }, + { + "Kind": "IFeaturizerInput", + "Settings": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel" + }, + { + "Name": "Data", + "Type": "DataView" + } + ] + }, + { + "Kind": "IMulticlassClassificationOutput", + "Settings": [] + }, + { + "Kind": "IRankingOutput", + "Settings": [] + }, + { + "Kind": "IRegressionOutput", + "Settings": [] + }, + { + "Kind": "ITrainerInput", + "Settings": [ + { + "Name": "TrainingData", + "Type": "DataView" + }, + { + "Name": "FeatureColumn", + "Type": "String" + } + ] + }, + { + "Kind": "ITrainerInputWithGroupId", + "Settings": [ + { + "Name": "GroupIdColumn", + "Type": "String" + }, + { + "Name": "WeightColumn", + "Type": "String" + }, + { + "Name": "LabelColumn", + "Type": "String" + }, + { + "Name": "TrainingData", + "Type": "DataView" + }, + { + "Name": "FeatureColumn", + "Type": "String" + } + ] + }, + { + "Kind": "ITrainerInputWithLabel", + "Settings": [ + { + "Name": "LabelColumn", + "Type": "String" + }, + { + "Name": "TrainingData", + "Type": "DataView" + }, + { + "Name": "FeatureColumn", + "Type": "String" + } + ] + }, + { + "Kind": "ITrainerInputWithWeight", + "Settings": [ + { + "Name": "WeightColumn", + "Type": "String" + }, + { + "Name": "LabelColumn", + "Type": "String" + }, + { + "Name": "TrainingData", + "Type": "DataView" + }, + { + "Name": "FeatureColumn", + "Type": "String" + } + ] + }, + { + "Kind": "ITrainerOutput", + "Settings": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel" + } + ] + }, + { + "Kind": "ITransformInput", + "Settings": [ + { + "Name": "Data", + "Type": "DataView" + } + ] + }, + { + "Kind": "ITransformOutput", + "Settings": [ + { + "Name": "OutputData", + "Type": "DataView" + }, + { + "Name": "Model", + "Type": "TransformModel" + } + ] + }, + { + "Kind": "IUnsupervisedTrainerWithWeight", + "Settings": [ + { + "Name": "WeightColumn", + "Type": "String" + }, + { + "Name": "TrainingData", + "Type": "DataView" + }, + { + "Name": "FeatureColumn", + "Type": "String" + } + ] + } + ] +} \ No newline at end of file From 65a5b416fedfb18a3894987cff7175e7a5418437 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Wed, 27 Mar 2024 12:08:34 -0700 Subject: [PATCH 14/26] Loosen tolerance in TestEntryPoints --- test/Microsoft.ML.Core.Tests/UnitTests/TestEntryPoints.cs | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/test/Microsoft.ML.Core.Tests/UnitTests/TestEntryPoints.cs b/test/Microsoft.ML.Core.Tests/UnitTests/TestEntryPoints.cs index 80acaf5d70..b86211f027 100644 --- a/test/Microsoft.ML.Core.Tests/UnitTests/TestEntryPoints.cs +++ b/test/Microsoft.ML.Core.Tests/UnitTests/TestEntryPoints.cs @@ -5208,7 +5208,7 @@ public void TestCrossValidationMacro() foldGetter(ref fold); Assert.True(ReadOnlyMemoryUtils.EqualsStr("Standard Deviation", fold)); if (w == 1) - Assert.Equal(1.585, stdev, .001); + Assert.Equal(1.585, stdev, .002); else Assert.Equal(1.39, stdev, 0.01); isWeightedGetter(ref isWeighted); From f7e20c4136054fee13e4e1fc795a9318e8840b02 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Wed, 27 Mar 2024 12:47:22 -0700 Subject: [PATCH 15/26] Add baseline file update capability --- test/Microsoft.ML.TestFramework/BaseTestBaseline.cs | 11 +++++++++++ 1 file changed, 11 insertions(+) diff --git a/test/Microsoft.ML.TestFramework/BaseTestBaseline.cs b/test/Microsoft.ML.TestFramework/BaseTestBaseline.cs index 3fce5c1bc2..9bcdbd884e 100644 --- a/test/Microsoft.ML.TestFramework/BaseTestBaseline.cs +++ b/test/Microsoft.ML.TestFramework/BaseTestBaseline.cs @@ -151,6 +151,12 @@ private IEnumerable GetConfigurationDirs() // This needs to come after win-arm but before win-x64 and win-x86 #if NETCOREAPP +#if NET9_0_OR_GREATER + configurationDirs.Add("net9.0"); +#endif +#if NET8_0_OR_GREATER + configurationDirs.Add("net8.0"); +#endif // Use netcoreapp result file if necessary. // The small difference comes from CPUMath using different instruction SSE (in CpuMathNative) vs // AVX, SSE or direct floating point calculation depending on hardware availability @@ -428,6 +434,11 @@ protected bool CheckEqualityCore(string dir, string name, string nameBase, bool bool res = CheckEqualityFromPathsCore(relPath, basePath, outPath, digitsOfPrecision: digitsOfPrecision, parseOption: parseOption); + if (!res && Environment.GetEnvironmentVariable("UpdateBaselines") == "1") + { + File.Copy(outPath, basePath, overwrite: true); + } + // No need to keep the raw (unnormalized) output file. if (normalize && res) File.Delete(outPath + RawSuffix); From af6b3b99939470864755fb9fb90aac63963bf64e Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Wed, 27 Mar 2024 12:47:52 -0700 Subject: [PATCH 16/26] Update baselines for net8.0 --- ...on-CV-breast-cancer.PAVcalibration-out.txt | 26 +- ...ron-CV-breast-cancer.PAVcalibration-rp.txt | 2 +- ...eptron-CV-breast-cancer.PAVcalibration.txt | 878 +++++------ ...n-CV-breast-cancer.calibrateRandom-out.txt | 34 +- ...on-CV-breast-cancer.calibrateRandom-rp.txt | 2 +- ...ptron-CV-breast-cancer.calibrateRandom.txt | 1022 ++++++------ ...ron-CV-breast-cancer.nocalibration-out.txt | 26 +- ...tron-CV-breast-cancer.nocalibration-rp.txt | 2 +- ...ceptron-CV-breast-cancer.nocalibration.txt | 874 +++++------ ...nTest-breast-cancer.PAVcalibration-out.txt | 26 +- ...inTest-breast-cancer.PAVcalibration-rp.txt | 2 +- ...TrainTest-breast-cancer.PAVcalibration.txt | 1366 ++++++++--------- ...Test-breast-cancer.calibrateRandom-out.txt | 26 +- ...nTest-breast-cancer.calibrateRandom-rp.txt | 2 +- ...rainTest-breast-cancer.calibrateRandom.txt | 1366 ++++++++--------- ...inTest-breast-cancer.nocalibration-out.txt | 18 +- ...ainTest-breast-cancer.nocalibration-rp.txt | 2 +- ...-TrainTest-breast-cancer.nocalibration.txt | 1366 ++++++++--------- .../LDSVM-def-CV-breast-cancer-out.txt | 48 +- .../netcoreapp/LDSVM-def-CV-breast-cancer.txt | 1366 ++++++++--------- .../LDSVM-nob-CV-breast-cancer-out.txt | 48 +- .../LDSVM-nob-CV-breast-cancer-rp.txt | 2 +- .../netcoreapp/LDSVM-nob-CV-breast-cancer.txt | 1366 ++++++++--------- .../LDSVM-nob-TrainTest-breast-cancer-out.txt | 34 +- .../LDSVM-nob-TrainTest-breast-cancer-rp.txt | 2 +- .../LDSVM-nob-TrainTest-breast-cancer.txt | 1366 ++++++++--------- ...gression-non-negative-CV-breast-cancer.txt | 1362 ++++++++-------- ...n-non-negative-TrainTest-breast-cancer.txt | 1360 ++++++++-------- ...ession-Non-Negative-TrainTest-iris-out.txt | 68 + ...ression-Non-Negative-TrainTest-iris-rp.txt | 4 + ...Regression-Non-Negative-TrainTest-iris.txt | 300 ++-- .../MulticlassLogisticRegression-CV-iris.txt | 300 ++-- ...classLogisticRegression-TrainTest-iris.txt | 300 ++-- .../netcoreapp/SymSGD-CV-breast-cancer.txt | 700 +++++++++ .../WE-AvgPer-TrainTest-breast-cancer-out.txt | 14 +- .../WE-AvgPer-TrainTest-breast-cancer-rp.txt | 2 +- .../WE-AvgPer-TrainTest-breast-cancer.txt | 1366 ++++++++--------- .../WE-Default-CV-breast-cancer.txt | 700 +++++++++ .../WE-Hetero-TrainTest-breast-cancer-out.txt | 118 ++ .../WE-Hetero-TrainTest-breast-cancer-rp.txt | 4 + .../WE-Hetero-TrainTest-breast-cancer.txt | 700 +++++++++ .../WE-Average-TrainTest-iris-out.txt | 8 +- .../WE-Average-TrainTest-iris-rp.txt | 2 +- .../netcoreapp/WE-Average-TrainTest-iris.txt | 300 ++-- .../WE-Bootstrap-TrainTest-iris-out.txt | 20 +- .../WE-Bootstrap-TrainTest-iris-rp.txt | 2 +- .../WE-Bootstrap-TrainTest-iris.txt | 300 ++-- .../WE-Stacking-TrainTest-iris-out.txt | 12 +- .../netcoreapp/WE-Stacking-TrainTest-iris.txt | 300 ++-- 49 files changed, 10904 insertions(+), 8610 deletions(-) create mode 100644 test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris-out.txt create mode 100644 test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris-rp.txt create mode 100644 test/BaselineOutput/Common/SymSGD/netcoreapp/SymSGD-CV-breast-cancer.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Default-CV-breast-cancer.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer-out.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer-rp.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer.txt diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration-out.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration-out.txt index 6da3dc0dcf..e598352377 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration-out.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration-out.txt @@ -13,15 +13,15 @@ Confusion table ||====================== PREDICTED || positive | negative | Recall TRUTH ||====================== - positive || 131 | 3 | 0.9776 - negative || 8 | 212 | 0.9636 + positive || 133 | 1 | 0.9925 + negative || 9 | 211 | 0.9591 ||====================== -Precision || 0.9424 | 0.9860 | -OVERALL 0/1 ACCURACY: 0.968927 +Precision || 0.9366 | 0.9953 | +OVERALL 0/1 ACCURACY: 0.971751 LOG LOSS/instance: Infinity Test-set entropy (prior Log-Loss/instance): 0.956998 LOG-LOSS REDUCTION (RIG): -Infinity -AUC: 0.994437 +AUC: 0.994403 Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. TEST POSITIVE RATIO: 0.3191 (105.0/(105.0+224.0)) Confusion table @@ -40,16 +40,16 @@ AUC: 0.997619 OVERALL RESULTS --------------------------------------- -AUC: 0.996028 (0.0016) -Accuracy: 0.972305 (0.0034) -Positive precision: 0.956660 (0.0142) -Positive recall: 0.964996 (0.0126) -Negative precision: 0.981961 (0.0041) -Negative recall: 0.975122 (0.0115) +AUC: 0.996011 (0.0016) +Accuracy: 0.973718 (0.0020) +Positive precision: 0.953747 (0.0171) +Positive recall: 0.972459 (0.0201) +Negative precision: 0.986580 (0.0087) +Negative recall: 0.972849 (0.0138) Log-loss: Infinity (NaN) Log-loss reduction: -Infinity (NaN) -F1 Score: 0.960623 (0.0009) -AUPRC: 0.992280 (0.0025) +F1 Score: 0.962653 (0.0011) +AUPRC: 0.992269 (0.0025) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration-rp.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration-rp.txt index 8dd72b1661..95bd831962 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration-rp.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration-rp.txt @@ -1,4 +1,4 @@ AveragedPerceptron AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings -0.996028 0.972305 0.95666 0.964996 0.981961 0.975122 Infinity -Infinity 0.960623 0.99228 AveragedPerceptron %Data% %Output% 99 0 0 maml.exe CV tr=AveragedPerceptron threads=- cali=PAV dout=%Output% data=%Data% seed=1 +0.996011 0.973718 0.953747 0.972459 0.98658 0.972849 Infinity -Infinity 0.962653 0.992269 AveragedPerceptron %Data% %Output% 99 0 0 maml.exe CV tr=AveragedPerceptron threads=- cali=PAV dout=%Output% data=%Data% seed=1 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration.txt index 1cef38aa8a..efa236c250 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.PAVcalibration.txt @@ -1,379 +1,379 @@ Instance Label Score Probability Log-loss Assigned -5 1 11.285979 1 -0 1 -6 0 -0.93471193 0.09090909 0.13750352804950119 0 -8 0 -3.7769966 1E-15 1.4415419267167138E-15 0 -9 0 -3.7947202 1E-15 1.4415419267167138E-15 0 -10 0 -4.7855167 1E-15 1.4415419267167138E-15 0 -11 0 -4.6251884 1E-15 1.4415419267167138E-15 0 -18 1 6.8803635 1 -0 1 -20 1 5.573552 1 -0 1 -21 1 6.7444105 1 -0 1 -25 1 1.2789736 0.875 0.19264507794239591 1 -28 0 -4.6251884 1E-15 1.4415419267167138E-15 0 -31 0 -4.3083706 1E-15 1.4415419267167138E-15 0 -32 1 6.9428844 1 -0 1 -35 0 -4.6251884 1E-15 1.4415419267167138E-15 0 -37 0 -1.8143315 0.09090909 0.13750352804950119 0 +5 1 11.925824 1 -0 1 +6 0 -0.4527979 0.2937258 0.50169967537103388 0 +8 0 -3.7962546 1E-15 1.4415419267167138E-15 0 +9 0 -3.8130417 1E-15 1.4415419267167138E-15 0 +10 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +11 0 -4.606969 1E-15 1.4415419267167138E-15 0 +18 1 7.3475924 1 -0 1 +20 1 6.1389017 1 -0 1 +21 1 7.1486177 1 -0 1 +25 1 1.6632223 0.88235295 0.18057223417631088 1 +28 0 -4.606969 1E-15 1.4415419267167138E-15 0 +31 0 -4.2645645 1E-15 1.4415419267167138E-15 0 +32 1 7.198575 1 -0 1 +35 0 -4.606969 1E-15 1.4415419267167138E-15 0 +37 0 -1.714282 0.1 0.15200309583369792 0 40 0 ? ? ? 0 -41 1 2.4075565 0.875 0.19264507794239591 1 -44 1 8.039285 1 -0 1 -45 0 -4.625085 1E-15 1.4415419267167138E-15 0 -46 1 5.138131 1 -0 1 -48 0 -3.4336777 1E-15 1.4415419267167138E-15 0 -50 1 2.7120514 0.875 0.19264507794239591 1 -51 1 -0.062075615 0.6666667 0.58496245772549416 0 -52 1 4.4027233 1 -0 1 -54 1 6.3079214 1 -0 1 -56 1 6.356518 1 -0 1 -60 1 1.9474735 0.875 0.19264507794239591 1 -63 1 0.78555584 0.6666667 0.58496245772549416 1 -64 0 -4.916337 1E-15 1.4415419267167138E-15 0 -66 0 -3.7260728 1E-15 1.4415419267167138E-15 0 -68 1 9.2772875 1 -0 1 -69 0 -4.4157114 1E-15 1.4415419267167138E-15 0 -70 0 -3.0868545 1E-15 1.4415419267167138E-15 0 -71 1 7.5159607 1 -0 1 -72 0 -1.8410158 0.09090909 0.13750352804950119 0 -73 1 7.1320066 1 -0 1 -74 1 2.4329157 0.875 0.19264507794239591 1 -76 0 -3.9190063 1E-15 1.4415419267167138E-15 0 -77 0 -3.1092033 1E-15 1.4415419267167138E-15 0 -79 0 -4.4391913 1E-15 1.4415419267167138E-15 0 -82 0 -3.1867537 1E-15 1.4415419267167138E-15 0 -88 0 -3.7260728 1E-15 1.4415419267167138E-15 0 -90 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -91 0 -4.5046597 1E-15 1.4415419267167138E-15 0 -92 0 -3.7260728 1E-15 1.4415419267167138E-15 0 -93 0 -4.916337 1E-15 1.4415419267167138E-15 0 -95 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -96 0 -4.7958083 1E-15 1.4415419267167138E-15 0 -97 0 -3.4349236 1E-15 1.4415419267167138E-15 0 -98 1 9.07517 1 -0 1 -99 1 8.952344 1 -0 1 -100 1 4.9092436 1 -0 1 -102 0 -3.393629 1E-15 1.4415419267167138E-15 0 -104 1 10.959611 1 -0 1 -105 1 2.0113592 0.875 0.19264507794239591 1 -106 1 8.251353 1 -0 1 -108 0 -4.5487204 1E-15 1.4415419267167138E-15 0 -109 1 5.864868 1 -0 1 -111 1 3.7846975 1 -0 1 -112 1 6.380026 1 -0 1 -113 1 9.461209 1 -0 1 -115 0 -3.6043515 1E-15 1.4415419267167138E-15 0 -117 1 7.9902315 1 -0 1 -120 0 -4.120878 1E-15 1.4415419267167138E-15 0 -121 0 -3.0707016 1E-15 1.4415419267167138E-15 0 -122 1 10.129083 1 -0 1 -123 1 4.173232 1 -0 1 -125 0 -4.916337 1E-15 1.4415419267167138E-15 0 -128 1 4.5026884 1 -0 1 -129 0 -3.7451797 1E-15 1.4415419267167138E-15 0 -131 0 -4.3083706 1E-15 1.4415419267167138E-15 0 -132 1 8.723828 1 -0 1 -133 0 -4.150231 1E-15 1.4415419267167138E-15 0 -137 0 -4.650857 1E-15 1.4415419267167138E-15 0 -138 0 -3.7104468 1E-15 1.4415419267167138E-15 0 -141 0 -4.942006 1E-15 1.4415419267167138E-15 0 -144 0 -4.6251884 1E-15 1.4415419267167138E-15 0 +41 1 2.5451756 0.88235295 0.18057223417631088 1 +44 1 8.165841 1 -0 1 +45 0 -4.602255 1E-15 1.4415419267167138E-15 0 +46 1 5.6216097 1 -0 1 +48 0 -3.379683 1E-15 1.4415419267167138E-15 0 +50 1 2.8003244 0.88235295 0.18057223417631088 1 +51 1 0.14775372 0.6666667 0.58496245772549416 1 +52 1 4.696246 1 -0 1 +54 1 6.743867 1 -0 1 +56 1 6.5947084 1 -0 1 +60 1 2.2064123 0.88235295 0.18057223417631088 1 +63 1 0.8789625 0.6666667 0.58496245772549416 1 +64 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +66 0 -3.697434 1E-15 1.4415419267167138E-15 0 +68 1 9.899808 1 -0 1 +69 0 -4.3595524 1E-15 1.4415419267167138E-15 0 +70 0 -3.0557137 1E-15 1.4415419267167138E-15 0 +71 1 7.555622 1 -0 1 +72 0 -1.676908 0.1 0.15200309583369792 0 +73 1 7.7111263 1 -0 1 +74 1 2.4994192 0.88235295 0.18057223417631088 1 +76 0 -3.839695 1E-15 1.4415419267167138E-15 0 +77 0 -3.1438046 1E-15 1.4415419267167138E-15 0 +79 0 -4.4265766 1E-15 1.4415419267167138E-15 0 +82 0 -3.1870723 1E-15 1.4415419267167138E-15 0 +88 0 -3.697434 1E-15 1.4415419267167138E-15 0 +90 0 -4.54813 1E-15 1.4415419267167138E-15 0 +91 0 -4.5069323 1E-15 1.4415419267167138E-15 0 +92 0 -3.697434 1E-15 1.4415419267167138E-15 0 +93 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +95 0 -4.54813 1E-15 1.4415419267167138E-15 0 +96 0 -4.790498 1E-15 1.4415419267167138E-15 0 +97 0 -3.413869 1E-15 1.4415419267167138E-15 0 +98 1 9.294541 1 -0 1 +99 1 9.62131 1 -0 1 +100 1 5.1074305 1 -0 1 +102 0 -3.3471546 1E-15 1.4415419267167138E-15 0 +104 1 11.120679 1 -0 1 +105 1 2.1430416 0.88235295 0.18057223417631088 1 +106 1 8.747506 1 -0 1 +108 0 -4.513379 1E-15 1.4415419267167138E-15 0 +109 1 6.3912535 1 -0 1 +111 1 4.1208715 0.88235295 0.18057223417631088 1 +112 1 7.006652 1 -0 1 +113 1 9.855809 1 -0 1 +115 0 -3.4127908 1E-15 1.4415419267167138E-15 0 +117 1 8.638749 1 -0 1 +120 0 -4.038975 1E-15 1.4415419267167138E-15 0 +121 0 -3.0156002 1E-15 1.4415419267167138E-15 0 +122 1 10.533511 1 -0 1 +123 1 4.6804914 1 -0 1 +125 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +128 1 4.9510326 1 -0 1 +129 0 -3.7898016 1E-15 1.4415419267167138E-15 0 +131 0 -4.2645645 1E-15 1.4415419267167138E-15 0 +132 1 9.206267 1 -0 1 +133 0 -4.134826 1E-15 1.4415419267167138E-15 0 +137 0 -4.6658077 1E-15 1.4415419267167138E-15 0 +138 0 -3.6895585 1E-15 1.4415419267167138E-15 0 +141 0 -4.9493732 1E-15 1.4415419267167138E-15 0 +144 0 -4.606969 1E-15 1.4415419267167138E-15 0 145 0 ? ? ? 0 -147 0 -4.3231397 1E-15 1.4415419267167138E-15 0 -150 0 -4.7855167 1E-15 1.4415419267167138E-15 0 -151 1 4.569235 1 -0 1 -152 1 8.551608 1 -0 1 -154 0 -5.233155 1E-15 1.4415419267167138E-15 0 -156 0 -4.3357234 1E-15 1.4415419267167138E-15 0 -161 0 -3.5422645 1E-15 1.4415419267167138E-15 0 +147 0 -4.255105 1E-15 1.4415419267167138E-15 0 +150 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +151 1 5.0748987 1 -0 1 +152 1 8.881612 1 -0 1 +154 0 -5.232939 1E-15 1.4415419267167138E-15 0 +156 0 -4.3134584 1E-15 1.4415419267167138E-15 0 +161 0 -3.5088563 1E-15 1.4415419267167138E-15 0 164 0 ? ? ? 0 -167 1 7.4310427 1 -0 1 -169 0 -5.2729545 1E-15 1.4415419267167138E-15 0 -171 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -173 1 13.79783 1 -0 1 -174 1 5.2489986 1 -0 1 -176 0 -4.3083706 1E-15 1.4415419267167138E-15 0 -177 1 6.069394 1 -0 1 -179 1 2.427494 0.875 0.19264507794239591 1 -180 0 -4.7855167 1E-15 1.4415419267167138E-15 0 -181 0 -5.233155 1E-15 1.4415419267167138E-15 0 -183 1 8.429484 1 -0 1 -187 1 11.391686 1 -0 1 -188 1 7.5789557 1 -0 1 -189 0 -3.686462 1E-15 1.4415419267167138E-15 0 -191 1 10.154208 1 -0 1 -192 0 -3.7517414 1E-15 1.4415419267167138E-15 0 -196 0 5.425536 1 Infinity 1 -198 0 -5.233155 1E-15 1.4415419267167138E-15 0 -199 0 -4.334039 1E-15 1.4415419267167138E-15 0 -201 1 9.260581 1 -0 1 -202 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -204 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -205 1 11.455533 1 -0 1 -206 1 6.5003185 1 -0 1 -207 0 -4.7855167 1E-15 1.4415419267167138E-15 0 -209 0 -3.6209207 1E-15 1.4415419267167138E-15 0 -210 1 12.897248 1 -0 1 -211 1 8.615066 1 -0 1 -212 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -216 0 -4.916337 1E-15 1.4415419267167138E-15 0 -218 1 7.456811 1 -0 1 -219 0 -2.4788876 1E-15 1.4415419267167138E-15 0 -223 1 4.949767 1 -0 1 -226 1 8.538112 1 -0 1 -228 0 -4.7855167 1E-15 1.4415419267167138E-15 0 -233 1 5.219512 1 -0 1 -237 1 6.374799 1 -0 1 -239 1 4.7197256 1 -0 1 -240 0 -2.0196419 0.09090909 0.13750352804950119 0 -241 0 -3.9108233 1E-15 1.4415419267167138E-15 0 -242 0 -4.3083706 1E-15 1.4415419267167138E-15 0 -244 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -246 1 10.46416 1 -0 1 -247 1 2.8055887 0.875 0.19264507794239591 1 -248 0 -2.8070307 1E-15 1.4415419267167138E-15 0 +167 1 7.490013 1 -0 1 +169 0 -5.254455 1E-15 1.4415419267167138E-15 0 +171 0 -4.54813 1E-15 1.4415419267167138E-15 0 +173 1 14.494651 1 -0 1 +174 1 5.7514915 1 -0 1 +176 0 -4.2645645 1E-15 1.4415419267167138E-15 0 +177 1 6.138485 1 -0 1 +179 1 2.8749352 0.88235295 0.18057223417631088 1 +180 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +181 0 -5.232939 1E-15 1.4415419267167138E-15 0 +183 1 9.06396 1 -0 1 +187 1 12.00492 1 -0 1 +188 1 8.132482 1 -0 1 +189 0 -3.6207743 1E-15 1.4415419267167138E-15 0 +191 1 10.797475 1 -0 1 +192 0 -3.7562728 1E-15 1.4415419267167138E-15 0 +196 0 5.8725023 1 Infinity 1 +198 0 -5.232939 1E-15 1.4415419267167138E-15 0 +199 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +201 1 9.770606 1 -0 1 +202 0 -4.54813 1E-15 1.4415419267167138E-15 0 +204 0 -4.54813 1E-15 1.4415419267167138E-15 0 +205 1 12.087651 1 -0 1 +206 1 6.641532 1 -0 1 +207 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +209 0 -3.5942607 1E-15 1.4415419267167138E-15 0 +210 1 13.547517 1 -0 1 +211 1 9.089206 1 -0 1 +212 0 -4.54813 1E-15 1.4415419267167138E-15 0 +216 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +218 1 7.7499723 1 -0 1 +219 0 -2.4297438 1E-15 1.4415419267167138E-15 0 +223 1 5.4305964 1 -0 1 +226 1 9.166205 1 -0 1 +228 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +233 1 5.6998796 1 -0 1 +237 1 6.476473 1 -0 1 +239 1 5.1697493 1 -0 1 +240 0 -1.9057703 0.1 0.15200309583369792 0 +241 0 -3.8436403 1E-15 1.4415419267167138E-15 0 +242 0 -4.2645645 1E-15 1.4415419267167138E-15 0 +244 0 -4.54813 1E-15 1.4415419267167138E-15 0 +246 1 11.068807 1 -0 1 +247 1 3.1855059 0.88235295 0.18057223417631088 1 +248 0 -2.7545462 1E-15 1.4415419267167138E-15 0 249 0 ? ? ? 0 -250 0 -4.652541 1E-15 1.4415419267167138E-15 0 -252 0 3.7707853 1 Infinity 1 -254 1 7.661195 1 -0 1 -257 0 -4.334039 1E-15 1.4415419267167138E-15 0 -258 0 -4.0172215 1E-15 1.4415419267167138E-15 0 -259 0 4.008581 1 Infinity 1 -260 1 7.7233286 1 -0 1 -262 1 9.050584 1 -0 1 -267 1 3.073165 0.875 0.19264507794239591 1 -268 1 8.364619 1 -0 1 -269 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -271 0 -3.4349236 1E-15 1.4415419267167138E-15 0 -272 1 3.073165 0.875 0.19264507794239591 1 +250 0 -4.655863 1E-15 1.4415419267167138E-15 0 +252 0 4.2659445 0.9602303 4.6521861799755815 1 +254 1 7.919627 1 -0 1 +257 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +258 0 -3.9809995 1E-15 1.4415419267167138E-15 0 +259 0 4.3407555 1 Infinity 1 +260 1 8.330071 1 -0 1 +262 1 9.629434 1 -0 1 +267 1 3.677929 0.88235295 0.18057223417631088 1 +268 1 8.609313 1 -0 1 +269 0 -4.54813 1E-15 1.4415419267167138E-15 0 +271 0 -3.413869 1E-15 1.4415419267167138E-15 0 +272 1 3.677929 0.88235295 0.18057223417631088 1 275 0 ? ? ? 0 -276 0 -4.334039 1E-15 1.4415419267167138E-15 0 -277 0 -4.916337 1E-15 1.4415419267167138E-15 0 -278 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -279 1 6.3361444 1 -0 1 -280 0 -4.0172215 1E-15 1.4415419267167138E-15 0 -283 1 5.09361 1 -0 1 -284 1 5.780157 1 -0 1 -285 1 12.663693 1 -0 1 -288 1 1.8098211 0.875 0.19264507794239591 1 -290 0 -5.233155 1E-15 1.4415419267167138E-15 0 -291 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -293 1 4.209258 1 -0 1 -296 0 1.7716074 0.875 3 1 +276 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +277 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +278 0 -4.54813 1E-15 1.4415419267167138E-15 0 +279 1 6.636298 1 -0 1 +280 0 -3.9809995 1E-15 1.4415419267167138E-15 0 +283 1 5.5930414 1 -0 1 +284 1 6.3683043 1 -0 1 +285 1 13.342931 1 -0 1 +288 1 2.133194 0.88235295 0.18057223417631088 1 +290 0 -5.232939 1E-15 1.4415419267167138E-15 0 +291 0 -4.54813 1E-15 1.4415419267167138E-15 0 +293 1 4.659646 1 -0 1 +296 0 2.0631151 0.88235295 3.0874629272416674 1 297 0 ? ? ? 0 -299 1 5.4912786 1 -0 1 -300 1 6.2749596 1 -0 1 -301 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -303 0 -4.5995197 1E-15 1.4415419267167138E-15 0 -304 1 4.320197 1 -0 1 -308 1 6.5411158 1 -0 1 -309 0 -1.7547832 0.09090909 0.13750352804950119 0 -311 0 -5.233155 1E-15 1.4415419267167138E-15 0 -312 1 3.2156925 0.875 0.19264507794239591 1 -314 0 -5.102334 1E-15 1.4415419267167138E-15 0 -316 1 3.9409618 1 -0 1 -317 1 8.260409 1 -0 1 -319 0 2.46204 0.875 3 1 +299 1 6.110527 1 -0 1 +300 1 6.6979456 1 -0 1 +301 0 -4.54813 1E-15 1.4415419267167138E-15 0 +303 0 -4.54813 1E-15 1.4415419267167138E-15 0 +304 1 4.865142 1 -0 1 +308 1 6.804652 1 -0 1 +309 0 -1.4760432 0.1 0.15200309583369792 0 +311 0 -5.232939 1E-15 1.4415419267167138E-15 0 +312 1 3.2489738 0.88235295 0.18057223417631088 1 +314 0 -5.070926 1E-15 1.4415419267167138E-15 0 +316 1 4.409379 1 -0 1 +317 1 8.798545 1 -0 1 +319 0 2.7989092 0.88235295 3.0874629272416674 1 321 0 ? ? ? 0 -323 1 4.269208 1 -0 1 -327 0 -4.916337 1E-15 1.4415419267167138E-15 0 -328 1 4.1030073 1 -0 1 -329 1 6.3562107 1 -0 1 -331 0 -3.1436715 1E-15 1.4415419267167138E-15 0 -332 0 -2.8411217 1E-15 1.4415419267167138E-15 0 -333 1 4.4240284 1 -0 1 -336 1 4.790291 1 -0 1 -338 0 -5.102334 1E-15 1.4415419267167138E-15 0 -343 0 -5.233155 1E-15 1.4415419267167138E-15 0 -344 1 8.780366 1 -0 1 -346 0 -2.788134 1E-15 1.4415419267167138E-15 0 -347 0 -5.0515347 1E-15 1.4415419267167138E-15 0 -348 1 -0.033994675 0.6666667 0.58496245772549416 0 -349 1 2.9944906 0.875 0.19264507794239591 1 -350 0 -3.776164 1E-15 1.4415419267167138E-15 0 -352 0 1.3297043 0.875 3 1 -353 1 8.744256 1 -0 1 -354 0 -4.916337 1E-15 1.4415419267167138E-15 0 -355 0 -3.6730852 1E-15 1.4415419267167138E-15 0 -358 1 6.1854076 1 -0 1 -360 1 14.4099455 1 -0 1 -361 1 6.113164 1 -0 1 -366 1 12.847375 1 -0 1 -368 0 -4.568268 1E-15 1.4415419267167138E-15 0 -370 0 -3.0281096 1E-15 1.4415419267167138E-15 0 -371 0 -4.568268 1E-15 1.4415419267167138E-15 0 -373 0 -3.605544 1E-15 1.4415419267167138E-15 0 -376 0 -4.916337 1E-15 1.4415419267167138E-15 0 -377 0 -5.102334 1E-15 1.4415419267167138E-15 0 -378 0 -3.5256414 1E-15 1.4415419267167138E-15 0 -379 0 -1.5692582 0.09090909 0.13750352804950119 0 -381 1 8.122036 1 -0 1 -383 0 -4.942006 1E-15 1.4415419267167138E-15 0 -384 0 -4.942006 1E-15 1.4415419267167138E-15 0 -387 0 -1.9415016 0.09090909 0.13750352804950119 0 -388 0 -4.44138 1E-15 1.4415419267167138E-15 0 -389 0 -2.9747353 1E-15 1.4415419267167138E-15 0 -391 1 8.779809 1 -0 1 -392 0 -4.334039 1E-15 1.4415419267167138E-15 0 -395 0 -4.334039 1E-15 1.4415419267167138E-15 0 -396 0 -4.0172215 1E-15 1.4415419267167138E-15 0 -398 0 -3.902061 1E-15 1.4415419267167138E-15 0 -399 0 -4.3200974 1E-15 1.4415419267167138E-15 0 -404 0 -4.508781 1E-15 1.4415419267167138E-15 0 -406 0 -3.4622765 1E-15 1.4415419267167138E-15 0 -409 0 -4.0015955 1E-15 1.4415419267167138E-15 0 -413 0 -3.1024804 1E-15 1.4415419267167138E-15 0 -414 1 5.959919 1 -0 1 -415 0 -0.721817 0.09090909 0.13750352804950119 0 -416 1 8.443301 1 -0 1 -418 0 -1.8258505 0.09090909 0.13750352804950119 0 -419 0 -3.8876748 1E-15 1.4415419267167138E-15 0 -422 0 -2.1972284 0.09090909 0.13750352804950119 0 -423 0 -3.0868545 1E-15 1.4415419267167138E-15 0 -428 0 -4.916337 1E-15 1.4415419267167138E-15 0 -429 0 -4.6251884 1E-15 1.4415419267167138E-15 0 -430 0 -4.2361884 1E-15 1.4415419267167138E-15 0 -434 0 5.330062 1 Infinity 1 -436 1 4.9601755 1 -0 1 -439 0 -4.068559 1E-15 1.4415419267167138E-15 0 -440 1 7.0005217 1 -0 1 -441 0 -1.8277493 0.09090909 0.13750352804950119 0 -442 0 -4.126358 1E-15 1.4415419267167138E-15 0 -449 1 9.384189 1 -0 1 -450 0 -3.936492 1E-15 1.4415419267167138E-15 0 -451 0 -4.068559 1E-15 1.4415419267167138E-15 0 -452 0 -4.358462 1E-15 1.4415419267167138E-15 0 -453 1 7.3491344 1 -0 1 -454 0 -4.259669 1E-15 1.4415419267167138E-15 0 -455 1 0.2950573 0.6666667 0.58496245772549416 1 -456 1 8.340758 1 -0 1 -457 1 7.996641 1 -0 1 -464 0 -4.3597083 1E-15 1.4415419267167138E-15 0 -465 1 8.680116 1 -0 1 -466 1 8.110646 1 -0 1 -467 1 6.858451 1 -0 1 -474 0 -4.068559 1E-15 1.4415419267167138E-15 0 -480 0 -4.254556 1E-15 1.4415419267167138E-15 0 -482 1 13.881022 1 -0 1 -483 1 9.617421 1 -0 1 -484 0 -3.736116 1E-15 1.4415419267167138E-15 0 -487 1 11.720016 1 -0 1 -489 1 -0.5905738 0.24775222 2.0130301133236248 0 -492 0 -4.228887 1E-15 1.4415419267167138E-15 0 -493 1 9.492114 1 -0 1 -495 0 -4.520036 1E-15 1.4415419267167138E-15 0 -497 0 -4.111538 1E-15 1.4415419267167138E-15 0 -501 0 -4.0428905 1E-15 1.4415419267167138E-15 0 -502 0 -3.8966928 1E-15 1.4415419267167138E-15 0 -504 0 -5.233155 1E-15 1.4415419267167138E-15 0 -507 0 -4.0115523 1E-15 1.4415419267167138E-15 0 -510 0 -5.233155 1E-15 1.4415419267167138E-15 0 -513 0 -4.520036 1E-15 1.4415419267167138E-15 0 -514 1 8.787938 1 -0 1 -517 0 -5.102334 1E-15 1.4415419267167138E-15 0 -519 1 6.320156 1 -0 1 -520 0 -5.0471582 1E-15 1.4415419267167138E-15 0 -521 0 -4.1737113 1E-15 1.4415419267167138E-15 0 -522 1 3.983387 1 -0 1 -523 1 6.156104 1 -0 1 -527 0 -3.7260728 1E-15 1.4415419267167138E-15 0 -528 0 -2.9663248 1E-15 1.4415419267167138E-15 0 -529 0 -4.228887 1E-15 1.4415419267167138E-15 0 -531 0 -3.4622765 1E-15 1.4415419267167138E-15 0 -532 0 -4.7855167 1E-15 1.4415419267167138E-15 0 -533 0 -4.334039 1E-15 1.4415419267167138E-15 0 -534 0 -4.6251884 1E-15 1.4415419267167138E-15 0 -535 0 -3.7884345 1E-15 1.4415419267167138E-15 0 -538 0 -4.0428905 1E-15 1.4415419267167138E-15 0 -539 0 -3.4605927 1E-15 1.4415419267167138E-15 0 -540 0 -3.3445406 1E-15 1.4415419267167138E-15 0 -541 0 -4.650857 1E-15 1.4415419267167138E-15 0 -544 0 -3.8141036 1E-15 1.4415419267167138E-15 0 -546 1 10.355874 1 -0 1 -547 0 -5.128003 1E-15 1.4415419267167138E-15 0 -548 0 -4.836854 1E-15 1.4415419267167138E-15 0 -549 1 5.2726183 1 -0 1 -557 0 -3.776164 1E-15 1.4415419267167138E-15 0 -558 0 -4.6251884 1E-15 1.4415419267167138E-15 0 -559 0 -3.7517414 1E-15 1.4415419267167138E-15 0 -560 0 -3.4349236 1E-15 1.4415419267167138E-15 0 -561 0 -3.4349236 1E-15 1.4415419267167138E-15 0 -563 0 -4.334039 1E-15 1.4415419267167138E-15 0 -565 1 10.456666 1 -0 1 -566 0 -3.6847782 1E-15 1.4415419267167138E-15 0 -569 1 8.855367 1 -0 1 -577 0 -4.916337 1E-15 1.4415419267167138E-15 0 -578 0 -4.916337 1E-15 1.4415419267167138E-15 0 -581 1 8.00238 1 -0 1 -582 1 7.645852 1 -0 1 -584 0 -3.1515741 1E-15 1.4415419267167138E-15 0 -586 1 12.260621 1 -0 1 -590 1 4.0090714 1 -0 1 -593 0 -3.736116 1E-15 1.4415419267167138E-15 0 -594 1 5.269803 1 -0 1 -600 0 -4.334039 1E-15 1.4415419267167138E-15 0 -602 0 -4.0428905 1E-15 1.4415419267167138E-15 0 -604 1 4.515382 1 -0 1 -606 0 -4.2135105 1E-15 1.4415419267167138E-15 0 -607 0 -5.233155 1E-15 1.4415419267167138E-15 0 -609 0 -4.068559 1E-15 1.4415419267167138E-15 0 -612 1 14.881892 1 -0 1 -613 0 -4.128848 1E-15 1.4415419267167138E-15 0 -614 0 -4.8111854 1E-15 1.4415419267167138E-15 0 +323 1 4.8855343 1 -0 1 +327 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +328 1 4.1213064 0.88235295 0.18057223417631088 1 +329 1 6.918497 1 -0 1 +331 0 -3.125589 1E-15 1.4415419267167138E-15 0 +332 0 -2.8317442 1E-15 1.4415419267167138E-15 0 +333 1 4.9879713 1 -0 1 +336 1 5.3119774 1 -0 1 +338 0 -5.070926 1E-15 1.4415419267167138E-15 0 +343 0 -5.232939 1E-15 1.4415419267167138E-15 0 +344 1 9.373975 1 -0 1 +346 0 -2.8051786 1E-15 1.4415419267167138E-15 0 +347 0 -5.0361757 1E-15 1.4415419267167138E-15 0 +348 1 0.09843922 0.6666667 0.58496245772549416 1 +349 1 3.543579 0.88235295 0.18057223417631088 1 +350 0 -3.7809267 1E-15 1.4415419267167138E-15 0 +352 0 1.5155859 0.88235295 3.0874629272416674 1 +353 1 9.391765 1 -0 1 +354 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +355 0 -3.6708684 1E-15 1.4415419267167138E-15 0 +358 1 6.2604885 1 -0 1 +360 1 15.035466 1 -0 1 +361 1 6.5571547 1 -0 1 +366 1 13.467439 1 -0 1 +368 0 -4.532379 1E-15 1.4415419267167138E-15 0 +370 0 -3.0555735 1E-15 1.4415419267167138E-15 0 +371 0 -4.532379 1E-15 1.4415419267167138E-15 0 +373 0 -3.5973978 1E-15 1.4415419267167138E-15 0 +376 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +377 0 -5.070926 1E-15 1.4415419267167138E-15 0 +378 0 -3.4778075 1E-15 1.4415419267167138E-15 0 +379 0 -1.5486526 0.1 0.15200309583369792 0 +381 1 8.643491 1 -0 1 +383 0 -4.9493732 1E-15 1.4415419267167138E-15 0 +384 0 -4.9493732 1E-15 1.4415419267167138E-15 0 +387 0 -1.9204488 0.1 0.15200309583369792 0 +388 0 -4.418391 1E-15 1.4415419267167138E-15 0 +389 0 -2.9322662 1E-15 1.4415419267167138E-15 0 +391 1 9.323798 1 -0 1 +392 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +395 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +396 0 -3.9809995 1E-15 1.4415419267167138E-15 0 +398 0 -3.9080298 1E-15 1.4415419267167138E-15 0 +399 0 -4.305583 1E-15 1.4415419267167138E-15 0 +404 0 -4.4998136 1E-15 1.4415419267167138E-15 0 +406 0 -3.4627619 1E-15 1.4415419267167138E-15 0 +409 0 -3.973124 1E-15 1.4415419267167138E-15 0 +413 0 -3.063589 1E-15 1.4415419267167138E-15 0 +414 1 6.5001116 1 -0 1 +415 0 -0.5379734 0.1598423 0.25126793919324036 0 +416 1 8.981729 1 -0 1 +418 0 -1.7460232 0.1 0.15200309583369792 0 +419 0 -3.9546041 1E-15 1.4415419267167138E-15 0 +422 0 -2.1661267 1E-15 1.4415419267167138E-15 0 +423 0 -3.0557137 1E-15 1.4415419267167138E-15 0 +428 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +429 0 -4.606969 1E-15 1.4415419267167138E-15 0 +430 0 -4.248364 1E-15 1.4415419267167138E-15 0 +434 0 5.678664 1 Infinity 1 +436 1 4.8968496 1 -0 1 +439 0 -4.098677 1E-15 1.4415419267167138E-15 0 +440 1 7.6677313 1 -0 1 +441 0 -1.6798639 0.1 0.15200309583369792 0 +442 0 -4.0051136 1E-15 1.4415419267167138E-15 0 +449 1 10.087396 1 -0 1 +450 0 -3.9024792 1E-15 1.4415419267167138E-15 0 +451 0 -4.098677 1E-15 1.4415419267167138E-15 0 +452 0 -4.348057 1E-15 1.4415419267167138E-15 0 +453 1 8.068194 1 -0 1 +454 0 -4.315388 1E-15 1.4415419267167138E-15 0 +455 1 0.4596901 0.6666667 0.58496245772549416 1 +456 1 8.942605 1 -0 1 +457 1 8.346686 1 -0 1 +464 0 -4.382242 1E-15 1.4415419267167138E-15 0 +465 1 8.953591 1 -0 1 +466 1 8.690221 1 -0 1 +467 1 7.4474783 1 -0 1 +474 0 -4.098677 1E-15 1.4415419267167138E-15 0 +480 0 -4.279069 1E-15 1.4415419267167138E-15 0 +482 1 14.295785 1 -0 1 +483 1 10.273198 1 -0 1 +484 0 -3.7483978 1E-15 1.4415419267167138E-15 0 +487 1 12.309564 1 -0 1 +489 1 -0.43852377 0.31616262 1.6612613055561321 0 +492 0 -4.22023 1E-15 1.4415419267167138E-15 0 +493 1 10.096398 1 -0 1 +495 0 -4.5037956 1E-15 1.4415419267167138E-15 0 +497 0 -4.155446 1E-15 1.4415419267167138E-15 0 +501 0 -4.039838 1E-15 1.4415419267167138E-15 0 +502 0 -3.8809624 1E-15 1.4415419267167138E-15 0 +504 0 -5.232939 1E-15 1.4415419267167138E-15 0 +507 0 -3.8678274 1E-15 1.4415419267167138E-15 0 +510 0 -5.232939 1E-15 1.4415419267167138E-15 0 +513 0 -4.5037956 1E-15 1.4415419267167138E-15 0 +514 1 9.508458 1 -0 1 +517 0 -5.070926 1E-15 1.4415419267167138E-15 0 +519 1 6.943595 1 -0 1 +520 0 -5.0525465 1E-15 1.4415419267167138E-15 0 +521 0 -4.2018504 1E-15 1.4415419267167138E-15 0 +522 1 4.3094034 0.9868349 0.019119380427994714 1 +523 1 6.7598047 1 -0 1 +527 0 -3.697434 1E-15 1.4415419267167138E-15 0 +528 0 -2.955677 1E-15 1.4415419267167138E-15 0 +529 0 -4.22023 1E-15 1.4415419267167138E-15 0 +531 0 -3.4627619 1E-15 1.4415419267167138E-15 0 +532 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +533 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +534 0 -4.606969 1E-15 1.4415419267167138E-15 0 +535 0 -3.6886954 1E-15 1.4415419267167138E-15 0 +538 0 -4.039838 1E-15 1.4415419267167138E-15 0 +539 0 -3.4727077 1E-15 1.4415419267167138E-15 0 +540 0 -3.3012362 1E-15 1.4415419267167138E-15 0 +541 0 -4.6658077 1E-15 1.4415419267167138E-15 0 +544 0 -3.7475338 1E-15 1.4415419267167138E-15 0 +546 1 11.016191 1 -0 1 +547 0 -5.1297655 1E-15 1.4415419267167138E-15 0 +548 0 -4.8462 1E-15 1.4415419267167138E-15 0 +549 1 5.5535192 1 -0 1 +557 0 -3.7809267 1E-15 1.4415419267167138E-15 0 +558 0 -4.606969 1E-15 1.4415419267167138E-15 0 +559 0 -3.7562728 1E-15 1.4415419267167138E-15 0 +560 0 -3.413869 1E-15 1.4415419267167138E-15 0 +561 0 -3.413869 1E-15 1.4415419267167138E-15 0 +563 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +565 1 11.259841 1 -0 1 +566 0 -3.6307197 1E-15 1.4415419267167138E-15 0 +569 1 9.242663 1 -0 1 +577 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +578 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +581 1 8.654239 1 -0 1 +582 1 8.128011 1 -0 1 +584 0 -3.123908 1E-15 1.4415419267167138E-15 0 +586 1 12.971318 1 -0 1 +590 1 4.1888075 0.9130087 0.13129950325655595 1 +593 0 -3.7483978 1E-15 1.4415419267167138E-15 0 +594 1 5.7850237 1 -0 1 +600 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +602 0 -4.039838 1E-15 1.4415419267167138E-15 0 +604 1 5.138941 1 -0 1 +606 0 -4.2233667 1E-15 1.4415419267167138E-15 0 +607 0 -5.232939 1E-15 1.4415419267167138E-15 0 +609 0 -4.098677 1E-15 1.4415419267167138E-15 0 +612 1 15.565835 1 -0 1 +613 0 -4.153376 1E-15 1.4415419267167138E-15 0 +614 0 -4.787361 1E-15 1.4415419267167138E-15 0 617 0 ? ? ? 0 -618 0 -4.0428905 1E-15 1.4415419267167138E-15 0 -619 0 -3.7517414 1E-15 1.4415419267167138E-15 0 -621 0 -0.14650154 0.6666667 1.5849625867124844 0 -622 0 -2.5446053 1E-15 1.4415419267167138E-15 0 -624 0 -3.7915406 1E-15 1.4415419267167138E-15 0 -627 0 -3.3132024 1E-15 1.4415419267167138E-15 0 -629 0 -4.3597083 1E-15 1.4415419267167138E-15 0 -633 1 4.0237722 1 -0 1 -634 0 -4.650857 1E-15 1.4415419267167138E-15 0 -638 0 -4.3597083 1E-15 1.4415419267167138E-15 0 -639 0 -3.776164 1E-15 1.4415419267167138E-15 0 -641 0 -4.334039 1E-15 1.4415419267167138E-15 0 -642 0 -4.334039 1E-15 1.4415419267167138E-15 0 -644 0 -4.942006 1E-15 1.4415419267167138E-15 0 -645 0 -4.334039 1E-15 1.4415419267167138E-15 0 -649 0 -4.334039 1E-15 1.4415419267167138E-15 0 -652 0 -3.567933 1E-15 1.4415419267167138E-15 0 -653 0 -4.0428905 1E-15 1.4415419267167138E-15 0 -654 0 -4.0172215 1E-15 1.4415419267167138E-15 0 -656 0 -3.7517414 1E-15 1.4415419267167138E-15 0 -657 0 0.674386 0.6666667 1.5849625867124844 1 -660 0 -4.916337 1E-15 1.4415419267167138E-15 0 -661 0 -3.7260728 1E-15 1.4415419267167138E-15 0 -665 0 -5.233155 1E-15 1.4415419267167138E-15 0 -668 1 3.2994661 0.875 0.19264507794239591 1 -670 1 6.4614477 1 -0 1 -678 0 -5.233155 1E-15 1.4415419267167138E-15 0 -679 0 -4.942006 1E-15 1.4415419267167138E-15 0 -680 1 14.404435 1 -0 1 -681 1 9.278363 1 -0 1 -682 0 -3.2511153 1E-15 1.4415419267167138E-15 0 -683 0 -5.233155 1E-15 1.4415419267167138E-15 0 -685 0 -5.233155 1E-15 1.4415419267167138E-15 0 -688 0 -4.3597083 1E-15 1.4415419267167138E-15 0 -689 0 -3.1943884 1E-15 1.4415419267167138E-15 0 -691 1 5.2444315 1 -0 1 -692 0 -4.650857 1E-15 1.4415419267167138E-15 0 -693 0 -4.042787 1E-15 1.4415419267167138E-15 0 -694 0 -4.057659 1E-15 1.4415419267167138E-15 0 -696 1 7.3569994 1 -0 1 -697 1 4.656295 1 -0 1 -698 1 5.6929607 1 -0 1 -0 0 -3.4721131 1E-15 1.4415419267167138E-15 0 +618 0 -4.039838 1E-15 1.4415419267167138E-15 0 +619 0 -3.7562728 1E-15 1.4415419267167138E-15 0 +621 0 0.0042414665 0.6666667 1.5849625867124844 1 +622 0 -2.5211115 1E-15 1.4415419267167138E-15 0 +624 0 -3.7777896 1E-15 1.4415419267167138E-15 0 +627 0 -3.1292257 1E-15 1.4415419267167138E-15 0 +629 0 -4.382242 1E-15 1.4415419267167138E-15 0 +633 1 4.158328 0.8943498 0.16108886142666864 1 +634 0 -4.6658077 1E-15 1.4415419267167138E-15 0 +638 0 -4.382242 1E-15 1.4415419267167138E-15 0 +639 0 -3.7809267 1E-15 1.4415419267167138E-15 0 +641 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +642 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +644 0 -4.9493732 1E-15 1.4415419267167138E-15 0 +645 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +649 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +652 0 -3.5676956 1E-15 1.4415419267167138E-15 0 +653 0 -4.039838 1E-15 1.4415419267167138E-15 0 +654 0 -3.9809995 1E-15 1.4415419267167138E-15 0 +656 0 -3.7562728 1E-15 1.4415419267167138E-15 0 +657 0 0.6928787 0.6666667 1.5849625867124844 1 +660 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +661 0 -3.697434 1E-15 1.4415419267167138E-15 0 +665 0 -5.232939 1E-15 1.4415419267167138E-15 0 +668 1 3.362691 0.88235295 0.18057223417631088 1 +670 1 6.9357576 1 -0 1 +678 0 -5.232939 1E-15 1.4415419267167138E-15 0 +679 0 -4.9493732 1E-15 1.4415419267167138E-15 0 +680 1 15.087735 1 -0 1 +681 1 9.870067 1 -0 1 +682 0 -3.2252913 1E-15 1.4415419267167138E-15 0 +683 0 -5.232939 1E-15 1.4415419267167138E-15 0 +685 0 -5.232939 1E-15 1.4415419267167138E-15 0 +688 0 -4.382242 1E-15 1.4415419267167138E-15 0 +689 0 -3.214981 1E-15 1.4415419267167138E-15 0 +691 1 5.504386 1 -0 1 +692 0 -4.6658077 1E-15 1.4415419267167138E-15 0 +693 0 -4.035124 1E-15 1.4415419267167138E-15 0 +694 0 -4.030379 1E-15 1.4415419267167138E-15 0 +696 1 7.6070995 1 -0 1 +697 1 5.0033865 1 -0 1 +698 1 6.1210947 1 -0 1 +0 0 -3.4721127 1E-15 1.4415419267167138E-15 0 1 0 2.4163914 0.9047619 3.3923175087700881 1 -2 0 -4.045404 1E-15 1.4415419267167138E-15 0 -3 0 2.9251795 0.9047619 3.3923175087700881 1 +2 0 -4.0454035 1E-15 1.4415419267167138E-15 0 +3 0 2.9251804 0.9047619 3.3923175087700881 1 4 0 -3.5088277 1E-15 1.4415419267167138E-15 0 -7 0 -4.670553 1E-15 1.4415419267167138E-15 0 -12 1 -0.34343147 0.51152503 0.96712324524188775 0 -13 0 -4.6186943 1E-15 1.4415419267167138E-15 0 +7 0 -4.6705537 1E-15 1.4415419267167138E-15 0 +12 1 -0.34343147 0.5115249 0.9671235814574376 0 +13 0 -4.618695 1E-15 1.4415419267167138E-15 0 14 1 7.360214 1 -0 1 -15 1 0.6494303 0.6 0.73696553683865695 1 +15 1 0.6494312 0.6 0.73696553683865695 1 16 0 -4.220706 1E-15 1.4415419267167138E-15 0 -17 0 -3.9551725 1E-15 1.4415419267167138E-15 0 -19 0 -2.9890537 1E-15 1.4415419267167138E-15 0 +17 0 -3.955172 1E-15 1.4415419267167138E-15 0 +19 0 -2.9890532 1E-15 1.4415419267167138E-15 0 22 0 -4.7037654 1E-15 1.4415419267167138E-15 0 23 1 ? ? ? 0 24 0 -5.4043503 1E-15 1.4415419267167138E-15 0 @@ -381,47 +381,47 @@ Instance Label Score Probability Log-loss Assigned 27 0 -3.7376466 1E-15 1.4415419267167138E-15 0 29 0 -5.4339433 1E-15 1.4415419267167138E-15 0 30 0 -4.649441 1E-15 1.4415419267167138E-15 0 -33 0 -4.698118 1E-15 1.4415419267167138E-15 0 +33 0 -4.6981187 1E-15 1.4415419267167138E-15 0 34 0 -4.4530277 1E-15 1.4415419267167138E-15 0 36 1 7.832773 1 -0 1 38 1 4.92861 0.93333334 0.09953566740867692 1 39 1 1.079258 0.6 0.73696553683865695 1 42 1 6.8985863 1 -0 1 -43 1 -0.49528694 0.5 1 0 +43 1 -0.49528742 0.5 1 0 47 0 -5.669884 1E-15 1.4415419267167138E-15 0 -49 1 5.3024044 0.93333334 0.09953566740867692 1 +49 1 5.3024054 0.93333334 0.09953566740867692 1 53 1 5.116103 0.93333334 0.09953566740867692 1 55 1 4.4195347 0.93333334 0.09953566740867692 1 -57 1 0.5701313 0.6 0.73696553683865695 1 -58 1 1.1371031 0.6 0.73696553683865695 1 +57 1 0.57013035 0.6 0.73696553683865695 1 +58 1 1.137104 0.6 0.73696553683865695 1 59 1 1.6442327 0.9047619 0.14438990028345636 1 61 0 -5.2770567 1E-15 1.4415419267167138E-15 0 -62 1 5.7670774 1 -0 1 -65 1 2.7867746 0.9047619 0.14438990028345636 1 +62 1 5.7670784 1 -0 1 +65 1 2.7867756 0.9047619 0.14438990028345636 1 67 1 2.9058514 0.9047619 0.14438990028345636 1 -75 0 -4.291042 1E-15 1.4415419267167138E-15 0 +75 0 -4.2910423 1E-15 1.4415419267167138E-15 0 78 0 -3.652576 1E-15 1.4415419267167138E-15 0 80 0 -3.2955709 1E-15 1.4415419267167138E-15 0 81 0 -3.9975338 1E-15 1.4415419267167138E-15 0 -83 0 -2.9832687 1E-15 1.4415419267167138E-15 0 -84 1 6.5824003 1 -0 1 +83 0 -2.9832683 1E-15 1.4415419267167138E-15 0 +84 1 6.5823994 1 -0 1 85 1 4.7604074 0.93333334 0.09953566740867692 1 86 1 1.466999 0.9047619 0.14438990028345636 1 -87 1 5.223543 0.93333334 0.09953566740867692 1 +87 1 5.223544 0.93333334 0.09953566740867692 1 89 0 -5.017977 1E-15 1.4415419267167138E-15 0 -94 0 -4.9212914 1E-15 1.4415419267167138E-15 0 -101 1 -0.85990286 0.41515666 1.2682722439451406 0 +94 0 -4.921291 1E-15 1.4415419267167138E-15 0 +101 1 -0.85990334 0.4151543 1.2682804255962932 0 103 1 0.36755466 0.6 0.73696553683865695 1 107 1 4.617798 0.93333334 0.09953566740867692 1 110 0 -3.1312823 1E-15 1.4415419267167138E-15 0 114 0 -2.8847933 1E-15 1.4415419267167138E-15 0 -116 0 -0.40737772 0.5 1 0 +116 0 -0.40737724 0.5 1 0 118 0 -5.2584124 1E-15 1.4415419267167138E-15 0 -119 0 -3.786881 1E-15 1.4415419267167138E-15 0 +119 0 -3.7868814 1E-15 1.4415419267167138E-15 0 124 1 5.0384197 0.93333334 0.09953566740867692 1 126 1 6.4765244 1 -0 1 127 0 -4.4382315 1E-15 1.4415419267167138E-15 0 -130 0 -3.2361722 1E-15 1.4415419267167138E-15 0 +130 0 -3.2361717 1E-15 1.4415419267167138E-15 0 134 0 -4.75809 1E-15 1.4415419267167138E-15 0 135 0 -2.670867 1E-15 1.4415419267167138E-15 0 136 0 -4.220706 1E-15 1.4415419267167138E-15 0 @@ -429,12 +429,12 @@ Instance Label Score Probability Log-loss Assigned 140 0 -4.9692993 1E-15 1.4415419267167138E-15 0 142 1 3.7043686 0.93333334 0.09953566740867692 1 143 0 -4.364107 1E-15 1.4415419267167138E-15 0 -146 1 0.36148834 0.6 0.73696553683865695 1 -148 0 -2.357305 1E-15 1.4415419267167138E-15 0 -149 1 8.756336 1 -0 1 +146 1 0.3614874 0.6 0.73696553683865695 1 +148 0 -2.3573046 1E-15 1.4415419267167138E-15 0 +149 1 8.756334 1 -0 1 153 0 -3.7005844 1E-15 1.4415419267167138E-15 0 155 1 3.1619148 0.9047619 0.14438990028345636 1 -157 0 -4.9212914 1E-15 1.4415419267167138E-15 0 +157 0 -4.921291 1E-15 1.4415419267167138E-15 0 158 0 ? ? ? 0 159 1 10.452139 1 -0 1 160 1 7.997595 1 -0 1 @@ -445,62 +445,62 @@ Instance Label Score Probability Log-loss Assigned 168 0 -4.4382315 1E-15 1.4415419267167138E-15 0 170 0 -4.9692993 1E-15 1.4415419267167138E-15 0 172 0 -5.669884 1E-15 1.4415419267167138E-15 0 -175 1 6.0381765 1 -0 1 -178 0 -3.9551725 1E-15 1.4415419267167138E-15 0 -182 0 -2.9890537 1E-15 1.4415419267167138E-15 0 +175 1 6.0381775 1 -0 1 +178 0 -3.955172 1E-15 1.4415419267167138E-15 0 +182 0 -2.9890532 1E-15 1.4415419267167138E-15 0 184 1 5.409273 1 -0 1 185 0 -4.8669662 1E-15 1.4415419267167138E-15 0 186 1 3.9876003 0.93333334 0.09953566740867692 1 190 1 10.521242 1 -0 1 193 0 -5.4043503 1E-15 1.4415419267167138E-15 0 194 0 -4.4382315 1E-15 1.4415419267167138E-15 0 -195 0 -3.9551725 1E-15 1.4415419267167138E-15 0 +195 0 -3.955172 1E-15 1.4415419267167138E-15 0 197 0 -2.8025956 1E-15 1.4415419267167138E-15 0 200 1 8.673521 1 -0 1 -203 0 -3.4721131 1E-15 1.4415419267167138E-15 0 -208 0 -5.350025 1E-15 1.4415419267167138E-15 0 +203 0 -3.4721127 1E-15 1.4415419267167138E-15 0 +208 0 -5.3500257 1E-15 1.4415419267167138E-15 0 213 1 12.300528 1 -0 1 214 1 11.895983 1 -0 1 -215 1 6.600219 1 -0 1 +215 1 6.6002197 1 -0 1 217 0 -5.4043503 1E-15 1.4415419267167138E-15 0 220 0 -5.181178 1E-15 1.4415419267167138E-15 0 221 1 7.9662914 1 -0 1 222 1 -2.1487255 1E-15 49.828921418077073 0 -224 1 8.4735565 1 -0 1 +224 1 8.473555 1 -0 1 225 0 -5.669884 1E-15 1.4415419267167138E-15 0 227 1 6.748211 1 -0 1 229 1 10.504805 1 -0 1 -230 1 4.829337 0.93333334 0.09953566740867692 1 +230 1 4.829338 0.93333334 0.09953566740867692 1 231 1 6.912092 1 -0 1 232 0 1.0722923 0.6 1.3219281808786905 1 234 0 -2.7037287 1E-15 1.4415419267167138E-15 0 235 0 ? ? ? 0 236 1 9.440506 1 -0 1 238 1 10.690645 1 -0 1 -243 0 -3.301972 1E-15 1.4415419267167138E-15 0 -245 0 -2.856113 1E-15 1.4415419267167138E-15 0 +243 0 -3.3019714 1E-15 1.4415419267167138E-15 0 +245 0 -2.8561125 1E-15 1.4415419267167138E-15 0 251 1 7.355525 1 -0 1 253 1 6.8985863 1 -0 1 -255 1 3.745204 0.93333334 0.09953566740867692 1 +255 1 3.745205 0.93333334 0.09953566740867692 1 256 0 -4.9692993 1E-15 1.4415419267167138E-15 0 261 1 9.009869 1 -0 1 263 1 7.1455708 1 -0 1 264 1 4.0145664 0.93333334 0.09953566740867692 1 -265 0 -2.5156913 1E-15 1.4415419267167138E-15 0 +265 0 -2.5156918 1E-15 1.4415419267167138E-15 0 266 1 7.325534 1 -0 1 270 1 5.5723915 1 -0 1 -273 1 0.03731823 0.6 0.73696553683865695 1 +273 1 0.037317276 0.6 0.73696553683865695 1 274 0 -4.2340226 1E-15 1.4415419267167138E-15 0 -281 0 -4.698118 1E-15 1.4415419267167138E-15 0 -282 1 2.860156 0.9047619 0.14438990028345636 1 +281 0 -4.6981187 1E-15 1.4415419267167138E-15 0 +282 1 2.860157 0.9047619 0.14438990028345636 1 286 1 12.544172 1 -0 1 287 0 -4.75809 1E-15 1.4415419267167138E-15 0 -289 1 6.6595697 1 -0 1 +289 1 6.6595707 1 -0 1 292 1 ? ? ? 0 294 0 ? ? ? 0 295 1 5.621522 1 -0 1 -298 0 -2.4584546 1E-15 1.4415419267167138E-15 0 -302 1 12.725584 1 -0 1 +298 0 -2.4584541 1E-15 1.4415419267167138E-15 0 +302 1 12.72558 1 -0 1 305 1 8.040865 1 -0 1 306 0 -5.4043503 1E-15 1.4415419267167138E-15 0 307 0 -5.4043503 1E-15 1.4415419267167138E-15 0 @@ -512,34 +512,34 @@ Instance Label Score Probability Log-loss Assigned 322 0 -4.4382315 1E-15 1.4415419267167138E-15 0 324 0 -5.4043503 1E-15 1.4415419267167138E-15 0 325 0 -3.7860875 1E-15 1.4415419267167138E-15 0 -326 1 3.6223288 0.90666085 0.14136509713640041 1 -330 1 4.9927454 0.93333334 0.09953566740867692 1 +326 1 3.6223297 0.9066624 0.14136263119498629 1 +330 1 4.9927444 0.93333334 0.09953566740867692 1 334 1 5.514736 1 -0 1 335 0 -5.935418 1E-15 1.4415419267167138E-15 0 337 0 -5.4043503 1E-15 1.4415419267167138E-15 0 339 1 5.1161976 0.93333334 0.09953566740867692 1 -340 1 5.5803347 1 -0 1 +340 1 5.5803356 1 -0 1 341 0 -5.4043503 1E-15 1.4415419267167138E-15 0 342 0 -5.4523587 1E-15 1.4415419267167138E-15 0 345 0 -5.935418 1E-15 1.4415419267167138E-15 0 -351 0 -4.9212914 1E-15 1.4415419267167138E-15 0 -356 1 -0.9127703 0.25679532 1.9613092019943366 0 +351 0 -4.921291 1E-15 1.4415419267167138E-15 0 +356 1 -0.9127703 0.2567953 1.9613093694259927 0 357 1 9.874601 1 -0 1 359 1 4.653248 0.93333334 0.09953566740867692 1 362 0 -3.5174994 1E-15 1.4415419267167138E-15 0 363 0 -1.6840982 1E-15 1.4415419267167138E-15 0 -364 0 -4.9212914 1E-15 1.4415419267167138E-15 0 +364 0 -4.921291 1E-15 1.4415419267167138E-15 0 365 0 -5.186825 1E-15 1.4415419267167138E-15 0 367 1 9.040358 1 -0 1 -369 0 -5.1140842 1E-15 1.4415419267167138E-15 0 +369 0 -5.1140847 1E-15 1.4415419267167138E-15 0 372 0 -3.9699688 1E-15 1.4415419267167138E-15 0 374 0 -4.4530277 1E-15 1.4415419267167138E-15 0 375 0 -5.935418 1E-15 1.4415419267167138E-15 0 380 0 -5.935418 1E-15 1.4415419267167138E-15 0 382 0 -3.5015903 1E-15 1.4415419267167138E-15 0 -385 0 -3.4273872 1E-15 1.4415419267167138E-15 0 +385 0 -3.4273868 1E-15 1.4415419267167138E-15 0 386 1 5.084236 0.93333334 0.09953566740867692 1 -390 0 -5.4467115 1E-15 1.4415419267167138E-15 0 +390 0 -5.446712 1E-15 1.4415419267167138E-15 0 393 0 -5.881093 1E-15 1.4415419267167138E-15 0 394 0 -4.909887 1E-15 1.4415419267167138E-15 0 397 0 -4.48624 1E-15 1.4415419267167138E-15 0 @@ -549,12 +549,12 @@ Instance Label Score Probability Log-loss Assigned 403 0 -3.7908158 1E-15 1.4415419267167138E-15 0 405 0 -5.669884 1E-15 1.4415419267167138E-15 0 407 0 -5.669884 1E-15 1.4415419267167138E-15 0 -408 0 -3.5375085 1E-15 1.4415419267167138E-15 0 +408 0 -3.537508 1E-15 1.4415419267167138E-15 0 410 0 -5.669884 1E-15 1.4415419267167138E-15 0 411 0 ? ? ? 0 -412 1 7.6394253 1 -0 1 +412 1 7.639426 1 -0 1 417 0 -5.669884 1E-15 1.4415419267167138E-15 0 -420 0 -2.6969714 1E-15 1.4415419267167138E-15 0 +420 0 -2.696971 1E-15 1.4415419267167138E-15 0 421 1 9.498289 1 -0 1 424 0 -4.9692993 1E-15 1.4415419267167138E-15 0 425 1 11.849485 1 -0 1 @@ -565,9 +565,9 @@ Instance Label Score Probability Log-loss Assigned 433 0 -4.013695 1E-15 1.4415419267167138E-15 0 435 1 7.0216722 1 -0 1 437 0 -4.48624 1E-15 1.4415419267167138E-15 0 -438 0 -3.5384212 1E-15 1.4415419267167138E-15 0 +438 0 -3.5384207 1E-15 1.4415419267167138E-15 0 443 0 -5.54259 1E-15 1.4415419267167138E-15 0 -444 0 -2.6761093 1E-15 1.4415419267167138E-15 0 +444 0 -2.6761103 1E-15 1.4415419267167138E-15 0 445 0 -5.4523587 1E-15 1.4415419267167138E-15 0 446 0 -5.935418 1E-15 1.4415419267167138E-15 0 447 0 -4.0031805 1E-15 1.4415419267167138E-15 0 @@ -576,31 +576,31 @@ Instance Label Score Probability Log-loss Assigned 459 0 -3.5017061 1E-15 1.4415419267167138E-15 0 460 0 -3.5568361 1E-15 1.4415419267167138E-15 0 461 0 -3.3979883 1E-15 1.4415419267167138E-15 0 -462 0 -3.0737762 1E-15 1.4415419267167138E-15 0 +462 0 -3.0737767 1E-15 1.4415419267167138E-15 0 463 0 -4.282031 1E-15 1.4415419267167138E-15 0 468 0 -4.48624 1E-15 1.4415419267167138E-15 0 469 0 -5.4191465 1E-15 1.4415419267167138E-15 0 470 0 -4.649441 1E-15 1.4415419267167138E-15 0 -471 0 -3.0737762 1E-15 1.4415419267167138E-15 0 +471 0 -3.0737767 1E-15 1.4415419267167138E-15 0 472 0 -3.4657965 1E-15 1.4415419267167138E-15 0 473 0 -4.48624 1E-15 1.4415419267167138E-15 0 475 0 -4.9692993 1E-15 1.4415419267167138E-15 0 -476 0 -4.2355022 1E-15 1.4415419267167138E-15 0 +476 0 -4.2355027 1E-15 1.4415419267167138E-15 0 477 0 -4.48624 1E-15 1.4415419267167138E-15 0 478 0 -3.744658 1E-15 1.4415419267167138E-15 0 479 1 6.673233 1 -0 1 481 0 -2.5658813 1E-15 1.4415419267167138E-15 0 485 0 -4.6490927 1E-15 1.4415419267167138E-15 0 486 0 -4.649441 1E-15 1.4415419267167138E-15 0 -488 1 0.9591036 0.6 0.73696553683865695 1 +488 1 0.95910263 0.6 0.73696553683865695 1 490 0 -5.935418 1E-15 1.4415419267167138E-15 0 -491 1 5.556223 1 -0 1 +491 1 5.556222 1 -0 1 494 0 -0.015696526 0.59426165 1.3013784104855217 0 496 0 -5.881093 1E-15 1.4415419267167138E-15 0 498 0 -4.220706 1E-15 1.4415419267167138E-15 0 499 0 -4.220706 1E-15 1.4415419267167138E-15 0 -500 0 -2.9890537 1E-15 1.4415419267167138E-15 0 -503 0 -3.9551725 1E-15 1.4415419267167138E-15 0 +500 0 -2.9890532 1E-15 1.4415419267167138E-15 0 +503 0 -3.955172 1E-15 1.4415419267167138E-15 0 505 0 -4.4522543 1E-15 1.4415419267167138E-15 0 506 1 8.766861 1 -0 1 508 0 -4.0031805 1E-15 1.4415419267167138E-15 0 @@ -609,12 +609,12 @@ Instance Label Score Probability Log-loss Assigned 512 0 -4.0031805 1E-15 1.4415419267167138E-15 0 515 1 7.3749876 1 -0 1 516 0 -5.881093 1E-15 1.4415419267167138E-15 0 -518 0 -4.49539 1E-15 1.4415419267167138E-15 0 +518 0 -4.4953895 1E-15 1.4415419267167138E-15 0 524 0 -4.7037654 1E-15 1.4415419267167138E-15 0 525 0 -4.746127 1E-15 1.4415419267167138E-15 0 526 0 -4.48624 1E-15 1.4415419267167138E-15 0 -530 1 5.325206 0.93333334 0.09953566740867692 1 -536 0 -3.4721131 1E-15 1.4415419267167138E-15 0 +530 1 5.325205 0.93333334 0.09953566740867692 1 +536 0 -3.4721127 1E-15 1.4415419267167138E-15 0 537 0 -3.2213755 1E-15 1.4415419267167138E-15 0 542 0 -3.4861355 1E-15 1.4415419267167138E-15 0 543 0 -4.220706 1E-15 1.4415419267167138E-15 0 @@ -622,7 +622,7 @@ Instance Label Score Probability Log-loss Assigned 550 0 -4.7037654 1E-15 1.4415419267167138E-15 0 551 0 -5.4043503 1E-15 1.4415419267167138E-15 0 552 0 -3.2360563 1E-15 1.4415419267167138E-15 0 -553 0 -1.4225526 1E-15 1.4415419267167138E-15 0 +553 0 -1.4225531 1E-15 1.4415419267167138E-15 0 554 0 -4.9692993 1E-15 1.4415419267167138E-15 0 555 0 -1.7431297 1E-15 1.4415419267167138E-15 0 556 0 -2.9508896 1E-15 1.4415419267167138E-15 0 @@ -631,7 +631,7 @@ Instance Label Score Probability Log-loss Assigned 567 0 -3.4350505 1E-15 1.4415419267167138E-15 0 568 1 3.5748348 0.9047619 0.14438990028345636 1 570 1 6.466878 1 -0 1 -571 1 9.048693 1 -0 1 +571 1 9.048691 1 -0 1 572 0 -4.7037654 1E-15 1.4415419267167138E-15 0 573 0 -5.669884 1E-15 1.4415419267167138E-15 0 574 1 5.533702 1 -0 1 @@ -644,15 +644,15 @@ Instance Label Score Probability Log-loss Assigned 587 0 -3.5334377 1E-15 1.4415419267167138E-15 0 588 1 4.6442137 0.93333334 0.09953566740867692 1 589 0 -4.0031805 1E-15 1.4415419267167138E-15 0 -591 1 4.243066 0.93333334 0.09953566740867692 1 -592 1 4.8517914 0.93333334 0.09953566740867692 1 +591 1 4.243067 0.93333334 0.09953566740867692 1 +592 1 4.8517895 0.93333334 0.09953566740867692 1 595 0 -3.7376466 1E-15 1.4415419267167138E-15 0 596 0 -3.9699688 1E-15 1.4415419267167138E-15 0 597 0 -2.9706383 1E-15 1.4415419267167138E-15 0 598 0 -4.7037654 1E-15 1.4415419267167138E-15 0 599 0 -2.9381208 1E-15 1.4415419267167138E-15 0 601 0 -5.6155596 1E-15 1.4415419267167138E-15 0 -603 1 3.1762495 0.9047619 0.14438990028345636 1 +603 1 3.1762505 0.9047619 0.14438990028345636 1 605 1 8.159748 1 -0 1 608 1 8.079367 1 -0 1 610 1 6.972576 1 -0 1 @@ -667,15 +667,15 @@ Instance Label Score Probability Log-loss Assigned 630 0 -2.7601237 1E-15 1.4415419267167138E-15 0 631 0 -3.7376466 1E-15 1.4415419267167138E-15 0 632 0 -5.935418 1E-15 1.4415419267167138E-15 0 -635 0 -4.217087 1E-15 1.4415419267167138E-15 0 -636 1 8.162584 1 -0 1 +635 0 -4.2170873 1E-15 1.4415419267167138E-15 0 +636 1 8.162586 1 -0 1 637 0 -2.370799 1E-15 1.4415419267167138E-15 0 -640 0 -4.0398955 1E-15 1.4415419267167138E-15 0 +640 0 -4.039895 1E-15 1.4415419267167138E-15 0 643 0 -5.935418 1E-15 1.4415419267167138E-15 0 646 0 -5.426158 1E-15 1.4415419267167138E-15 0 647 0 -5.4890733 1E-15 1.4415419267167138E-15 0 -648 1 8.579458 1 -0 1 -650 0 -3.6219687 1E-15 1.4415419267167138E-15 0 +648 1 8.579456 1 -0 1 +650 0 -3.6219683 1E-15 1.4415419267167138E-15 0 651 0 -4.965017 1E-15 1.4415419267167138E-15 0 655 0 -4.7037654 1E-15 1.4415419267167138E-15 0 658 1 7.546403 1 -0 1 @@ -683,11 +683,11 @@ Instance Label Score Probability Log-loss Assigned 662 0 -5.1684093 1E-15 1.4415419267167138E-15 0 663 0 -5.1684093 1E-15 1.4415419267167138E-15 0 664 0 -4.2574205 1E-15 1.4415419267167138E-15 0 -666 0 -3.070212 1E-15 1.4415419267167138E-15 0 +666 0 -3.0702114 1E-15 1.4415419267167138E-15 0 667 0 -4.4382315 1E-15 1.4415419267167138E-15 0 -669 1 6.981786 1 -0 1 +669 1 6.9817867 1 -0 1 671 0 -3.9565368 1E-15 1.4415419267167138E-15 0 -672 0 -4.9212914 1E-15 1.4415419267167138E-15 0 +672 0 -4.921291 1E-15 1.4415419267167138E-15 0 673 0 -3.289723 1E-15 1.4415419267167138E-15 0 674 0 -5.669884 1E-15 1.4415419267167138E-15 0 675 0 -3.7340279 1E-15 1.4415419267167138E-15 0 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom-out.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom-out.txt index bd0288a2bb..7d3ac2552a 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom-out.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom-out.txt @@ -11,15 +11,15 @@ Confusion table ||====================== PREDICTED || positive | negative | Recall TRUTH ||====================== - positive || 131 | 3 | 0.9776 - negative || 8 | 212 | 0.9636 + positive || 133 | 1 | 0.9925 + negative || 9 | 211 | 0.9591 ||====================== -Precision || 0.9424 | 0.9860 | -OVERALL 0/1 ACCURACY: 0.968927 -LOG LOSS/instance: 0.138699 +Precision || 0.9366 | 0.9953 | +OVERALL 0/1 ACCURACY: 0.971751 +LOG LOSS/instance: 0.139629 Test-set entropy (prior Log-Loss/instance): 0.956998 -LOG-LOSS REDUCTION (RIG): 0.855069 -AUC: 0.994437 +LOG-LOSS REDUCTION (RIG): 0.854097 +AUC: 0.994403 Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. TEST POSITIVE RATIO: 0.3191 (105.0/(105.0+224.0)) Confusion table @@ -38,16 +38,16 @@ AUC: 0.997619 OVERALL RESULTS --------------------------------------- -AUC: 0.996028 (0.0016) -Accuracy: 0.972305 (0.0034) -Positive precision: 0.956660 (0.0142) -Positive recall: 0.964996 (0.0126) -Negative precision: 0.981961 (0.0041) -Negative recall: 0.975122 (0.0115) -Log-loss: 0.129850 (0.0088) -Log-loss reduction: 0.860569 (0.0055) -F1 Score: 0.960623 (0.0009) -AUPRC: 0.992280 (0.0025) +AUC: 0.996011 (0.0016) +Accuracy: 0.973718 (0.0020) +Positive precision: 0.953747 (0.0171) +Positive recall: 0.972459 (0.0201) +Negative precision: 0.986580 (0.0087) +Negative recall: 0.972849 (0.0138) +Log-loss: 0.130315 (0.0093) +Log-loss reduction: 0.860083 (0.0060) +F1 Score: 0.962653 (0.0011) +AUPRC: 0.992269 (0.0025) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom-rp.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom-rp.txt index e59aeb877d..ae8060865c 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom-rp.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom-rp.txt @@ -1,4 +1,4 @@ AveragedPerceptron AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings -0.996028 0.972305 0.95666 0.964996 0.981961 0.975122 0.12985 0.860569 0.960623 0.99228 AveragedPerceptron %Data% %Output% 99 0 0 maml.exe CV tr=AveragedPerceptron threads=- numcali=200 dout=%Output% data=%Data% seed=1 +0.996011 0.973718 0.953747 0.972459 0.98658 0.972849 0.130315 0.860083 0.962653 0.992269 AveragedPerceptron %Data% %Output% 99 0 0 maml.exe CV tr=AveragedPerceptron threads=- numcali=200 dout=%Output% data=%Data% seed=1 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom.txt index 2a3a57c76c..5dfec709ea 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.calibrateRandom.txt @@ -1,506 +1,506 @@ Instance Label Score Probability Log-loss Assigned -5 1 11.285979 0.999881 0.00017164889637290732 1 -6 0 -0.93471193 0.23396653 0.38452066506363075 0 -8 0 -3.7769966 0.027554922 0.040311323096438006 0 -9 0 -3.7947202 0.027160425 0.039726175859896601 0 -10 0 -4.7855167 0.012041707 0.01747795506449706 0 -11 0 -4.6251884 0.013746234 0.019969190332632794 0 -18 1 6.8803635 0.9952804 0.0068250837128981312 1 -20 1 5.573552 0.9860489 0.020268933879564289 1 -21 1 6.7444105 0.994715 0.0076448984232590942 1 -25 1 1.2789736 0.6605421 0.59827764318621834 1 -28 0 -4.6251884 0.013746234 0.019969190332632794 0 -31 0 -4.3083706 0.017843187 0.025974708346612547 0 -32 1 6.9428844 0.9955198 0.0064780602842336027 1 -35 0 -4.6251884 0.013746234 0.019969190332632794 0 -37 0 -1.8143315 0.127655 0.19702928089143074 0 +5 1 11.925824 0.99988866 0.00016064073906035849 1 +6 0 -0.4527979 0.28799012 0.49003084056125074 0 +8 0 -3.7962546 0.026382556 0.038573078561515095 0 +9 0 -3.8130417 0.026036168 0.038059895132779242 0 +10 0 -4.7285223 0.0125917215 0.018281354087045291 0 +11 0 -4.606969 0.013873975 0.020156062560767365 0 +18 1 7.3475924 0.99550986 0.006492485535140746 1 +20 1 6.1389017 0.9881576 0.017186984398807038 1 +21 1 7.1486177 0.9947303 0.0076226814052998944 1 +25 1 1.6632223 0.6911742 0.53287870995530984 1 +28 0 -4.606969 0.013873975 0.020156062560767365 0 +31 0 -4.2645645 0.01821836 0.026525908568759592 0 +32 1 7.198575 0.99493784 0.0073217047835288053 1 +35 0 -4.606969 0.013873975 0.020156062560767365 0 +37 0 -1.714282 0.1272981 0.19643915971313131 0 40 0 ? ? ? 0 -41 1 2.4075565 0.8333802 0.26295323230625189 1 -44 1 8.039285 0.99820465 0.0025924726101035612 1 -45 0 -4.625085 0.013747407 0.019970906883975219 0 -46 1 5.138131 0.980041 0.029085949424629401 1 -48 0 -3.4336777 0.03638818 0.053476006321589616 0 -50 1 2.7120514 0.8658191 0.20786247380562253 1 -51 1 -0.062075615 0.38791537 1.366186144764757 0 -52 1 4.4027233 0.9636923 0.053355505841607219 1 -54 1 6.3079214 0.99240357 0.011001174399811285 1 -56 1 6.356518 0.992704 0.010564526737643879 1 -60 1 1.9474735 0.77292717 0.37159562329703083 1 -63 1 0.78555584 0.562906 0.82903400016193718 1 -64 0 -4.916337 0.010806993 0.015676053599848205 0 -66 0 -3.7260728 0.028719628 0.042040287835734616 0 -68 1 9.2772875 0.9993619 0.00092091699247825402 1 -69 0 -4.4157114 0.016335884 0.023762321052263443 0 -70 0 -3.0868545 0.04804749 0.071038491836753939 0 -71 1 7.5159607 0.9972213 0.0040144096248406758 1 -72 0 -1.8410158 0.12518987 0.19295816947573857 0 -73 1 7.1320066 0.99617285 0.0055320097082099612 1 -74 1 2.4329157 0.836305 0.2578988681156526 1 -76 0 -3.9190063 0.024544321 0.035851771146192368 0 -77 0 -3.1092033 0.047199592 0.069754063898082266 0 -79 0 -4.4391913 0.01602325 0.023303868229407783 0 -82 0 -3.1867537 0.04436643 0.065470561192308124 0 -88 0 -3.7260728 0.028719628 0.042040287835734616 0 -90 0 -4.5995197 0.014040394 0.020399552811077585 0 -91 0 -4.5046597 0.015182304 0.022071409609242969 0 -92 0 -3.7260728 0.028719628 0.042040287835734616 0 -93 0 -4.916337 0.010806993 0.015676053599848205 0 -95 0 -4.5995197 0.014040394 0.020399552811077585 0 -96 0 -4.7958083 0.011939717 0.017329030128849933 0 -97 0 -3.4349236 0.036351655 0.053421320856594394 0 -98 1 9.07517 0.9992444 0.0010905240888066275 1 -99 1 8.952344 0.99916273 0.0012084261062733697 1 -100 1 4.9092436 0.9759306 0.035149577872267759 1 -102 0 -3.393629 0.037581276 0.055263384475561735 0 -104 1 10.959611 0.9998437 0.00022548687566550783 1 -105 1 2.0113592 0.7821696 0.35444666591631291 1 -106 1 8.251353 0.99849606 0.002171366339988838 1 -108 0 -4.5487204 0.01464096 0.021278592679852121 0 -109 1 5.864868 0.989033 0.015909459906802714 1 -111 1 3.7846975 0.9405741 0.088386474079630892 1 -112 1 6.380026 0.992845 0.01035959049172811 1 -113 1 9.461209 0.9994528 0.00078961641462542323 1 -115 0 -3.6043515 0.031700347 0.046474516862521159 0 -117 1 7.9902315 0.99812955 0.0027010206381427706 1 -120 0 -4.120878 0.020810021 0.030339301406693928 0 -121 0 -3.0707016 0.0486693 0.071981160833770388 0 -122 1 10.129083 0.99968696 0.00045169714530399755 1 -123 1 4.173232 0.95634377 0.064398788225312462 1 -125 0 -4.916337 0.010806993 0.015676053599848205 0 -128 1 4.5026884 0.9665074 0.049147353467874903 1 -129 0 -3.7451797 0.028277127 0.041383166713241507 0 -131 0 -4.3083706 0.017843187 0.025974708346612547 0 -132 1 8.723828 0.99898654 0.001462851900897584 1 -133 0 -4.150231 0.020315535 0.029610931995403426 0 -137 0 -4.650857 0.013458151 0.019547844212221475 0 -138 0 -3.7104468 0.0290865 0.042585325778461289 0 -141 0 -4.942006 0.010579844 0.015344805197525327 0 -144 0 -4.6251884 0.013746234 0.019969190332632794 0 +41 1 2.5451756 0.8203417 0.28570311776320229 1 +44 1 8.165841 0.9976778 0.0033541180091675089 1 +45 0 -4.602255 0.0139262155 0.020232492307431734 0 +46 1 5.6216097 0.9821182 0.026031444169057363 1 +48 0 -3.379683 0.03656119 0.053735055635659848 0 +50 1 2.8003244 0.8487653 0.2365623958486662 1 +51 1 0.14775372 0.39660627 1.3342206221239763 1 +52 1 4.696246 0.9629518 0.054464539198696107 1 +54 1 6.743867 0.9927051 0.01056288089528439 1 +56 1 6.5947084 0.99177784 0.011911108302460895 1 +60 1 2.2064123 0.7763946 0.36513800059720591 1 +63 1 0.8789625 0.5427823 0.88155440156950038 1 +64 0 -4.8905344 0.011062949 0.016049403073119514 0 +66 0 -3.697434 0.028514326 0.041735373036084926 0 +68 1 9.899808 0.99942744 0.00082626923994757486 1 +69 0 -4.3595524 0.016894376 0.024581667974386916 0 +70 0 -3.0557137 0.046994247 0.069443170930934078 0 +71 1 7.555622 0.9962023 0.0054893674226172791 1 +72 0 -1.676908 0.13069293 0.20206221530042301 0 +73 1 7.7111263 0.99664944 0.0048419465200441776 1 +74 1 2.4994192 0.8148248 0.29543816942188406 1 +76 0 -3.839695 0.025495268 0.037258904547537594 0 +77 0 -3.1438046 0.043905534 0.064774925308085776 0 +79 0 -4.4265766 0.016017534 0.023295486850883066 0 +82 0 -3.1870723 0.042460293 0.062595780681733368 0 +88 0 -3.697434 0.028514326 0.041735373036084926 0 +90 0 -4.54813 0.01454009 0.021130913771998817 0 +91 0 -4.5069323 0.015025147 0.021841203066639781 0 +92 0 -3.697434 0.028514326 0.041735373036084926 0 +93 0 -4.8905344 0.011062949 0.016049403073119514 0 +95 0 -4.54813 0.01454009 0.021130913771998817 0 +96 0 -4.790498 0.011983715 0.017393273625651583 0 +97 0 -3.413869 0.0356 0.052296442645105176 0 +98 1 9.294541 0.99906635 0.0013475973096922743 1 +99 1 9.62131 0.99928296 0.001034846704541113 1 +100 1 5.1074305 0.97314864 0.039267908030735391 1 +102 0 -3.3471546 0.037498932 0.055139953861279192 0 +104 1 11.120679 0.99978656 0.00030796780451509731 1 +105 1 2.1430416 0.7673742 0.38199780127191252 1 +106 1 8.747506 0.99854773 0.0020967014982888315 1 +108 0 -4.513379 0.014948205 0.021728510121695226 0 +109 1 6.3912535 0.99032205 0.014030327944048709 1 +111 1 4.1208715 0.9422745 0.085780676489496735 1 +112 1 7.006652 0.9940932 0.0085470080951718218 1 +113 1 9.855809 0.9994067 0.00085621167564539895 1 +115 0 -3.4127908 0.035629936 0.052341226773579816 0 +117 1 8.638749 0.99841446 0.0022892705964363013 1 +120 0 -4.038975 0.021784056 0.031775115415170981 0 +121 0 -3.0156002 0.048468232 0.071676271758961241 0 +122 1 10.533511 0.9996569 0.00049505099616089013 1 +123 1 4.6804914 0.9624947 0.055149541356901512 1 +125 0 -4.8905344 0.011062949 0.016049403073119514 0 +128 1 4.9510326 0.9696393 0.044479919754725615 1 +129 0 -3.7898016 0.0265169 0.038772160613755033 0 +131 0 -4.2645645 0.01821836 0.026525908568759592 0 +132 1 9.206267 0.99899733 0.0014472717652274012 1 +133 0 -4.134826 0.02019246 0.029429699984409106 0 +137 0 -4.6658077 0.013237966 0.019225886685070071 0 +138 0 -3.6895585 0.028691236 0.041998115839947503 0 +141 0 -4.9493732 0.010554424 0.015307739563326032 0 +144 0 -4.606969 0.013873975 0.020156062560767365 0 145 0 ? ? ? 0 -147 0 -4.3231397 0.017627962 0.025658598019937556 0 -150 0 -4.7855167 0.012041707 0.01747795506449706 0 -151 1 4.569235 0.96826327 0.046528728712941197 1 -152 1 8.551608 0.99882966 0.001689428468282927 1 -154 0 -5.233155 0.008311939 0.012041708383174634 0 -156 0 -4.3357234 0.017446598 0.025392274638807534 0 -161 0 -3.5422645 0.033333912 0.048910464842286831 0 +147 0 -4.255105 0.01835566 0.026727679372009493 0 +150 0 -4.7285223 0.0125917215 0.018281354087045291 0 +151 1 5.0748987 0.97245276 0.040299926989301821 1 +152 1 8.881612 0.99869674 0.0018814265868558253 1 +154 0 -5.232939 0.00841024 0.012184721289626915 0 +156 0 -4.3134584 0.017524607 0.025506821392282504 0 +161 0 -3.5088563 0.03305538 0.048494830542210592 0 164 0 ? ? ? 0 -167 1 7.4310427 0.9970173 0.00430952206969425 1 -169 0 -5.2729545 0.008041957 0.011648995447973737 0 -171 0 -4.5995197 0.014040394 0.020399552811077585 0 -173 1 13.79783 0.99998546 2.0982035982102257E-05 1 -174 1 5.2489986 0.9817766 0.026533320639140424 1 -176 0 -4.3083706 0.017843187 0.025974708346612547 0 -177 1 6.069394 0.99074155 0.013419335993965349 1 -179 1 2.427494 0.8356832 0.25897191578675427 1 -180 0 -4.7855167 0.012041707 0.01747795506449706 0 -181 0 -5.233155 0.008311939 0.012041708383174634 0 -183 1 8.429484 0.99870396 0.0018710080960925083 1 -187 1 11.391686 0.9998911 0.00015711470643243279 1 -188 1 7.5789557 0.9973635 0.0038086772809973425 1 -189 0 -3.686462 0.029658489 0.04343550262275462 0 -191 1 10.154208 0.99969345 0.00044232118620536842 1 -192 0 -3.7517414 0.028126696 0.041159842562845529 0 -196 0 5.425536 0.984239 5.9874954870425023 1 -198 0 -5.233155 0.008311939 0.012041708383174634 0 -199 0 -4.334039 0.017470764 0.025427758243753987 0 -201 1 9.260581 0.9993529 0.00093391003230289336 1 -202 0 -4.5995197 0.014040394 0.020399552811077585 0 -204 0 -4.5995197 0.014040394 0.020399552811077585 0 -205 1 11.455533 0.99989676 0.00014894466394889936 1 -206 1 6.5003185 0.9935255 0.0093710911166817908 1 -207 0 -4.7855167 0.012041707 0.01747795506449706 0 -209 0 -3.6209207 0.031277653 0.045844871724036791 0 -210 1 12.897248 0.99996907 4.4630188215668621E-05 1 -211 1 8.615066 0.9988901 0.0016021337380065527 1 -212 0 -4.5995197 0.014040394 0.020399552811077585 0 -216 0 -4.916337 0.010806993 0.015676053599848205 0 -218 1 7.456811 0.99708074 0.0042177565023021743 1 -219 0 -2.4788876 0.07743221 0.11627316752966073 0 -223 1 4.949767 0.97671413 0.033991720509423423 1 -226 1 8.538112 0.99881643 0.0017085410371302549 1 -228 0 -4.7855167 0.012041707 0.01747795506449706 0 -233 1 5.219512 0.98133 0.027189762853290386 1 -237 1 6.374799 0.9928139 0.01040480215647019 1 -239 1 4.7197256 0.9719118 0.04110271604515562 1 -240 0 -2.0196419 0.10972074 0.16767014249243722 0 -241 0 -3.9108233 0.02470874 0.036094967295392681 0 -242 0 -4.3083706 0.017843187 0.025974708346612547 0 -244 0 -4.5995197 0.014040394 0.020399552811077585 0 -246 1 10.46416 0.9997635 0.00034125393609690753 1 -247 1 2.8055887 0.8746519 0.19321912280766068 1 -248 0 -2.8070307 0.05995963 0.089205382448756254 0 +167 1 7.490013 0.99599624 0.0057878035564347926 1 +169 0 -5.254455 0.0082664015 0.011975461881993393 0 +171 0 -4.54813 0.01454009 0.021130913771998817 0 +173 1 14.494651 0.99998605 2.0122110477710732E-05 1 +174 1 5.7514915 0.983872 0.023457464452347754 1 +176 0 -4.2645645 0.01821836 0.026525908568759592 0 +177 1 6.138485 0.98815364 0.017192727854047185 1 +179 1 2.8749352 0.8563466 0.22373324768939593 1 +180 0 -4.7285223 0.0125917215 0.018281354087045291 0 +181 0 -5.232939 0.00841024 0.012184721289626915 0 +183 1 9.06396 0.9988752 0.0016236556168074772 1 +187 1 12.00492 0.9998956 0.00015066466904787792 1 +188 1 8.132482 0.99761444 0.0034457424608055095 1 +189 0 -3.6207743 0.030282307 0.044363288606634702 0 +191 1 10.797475 0.99972284 0.00039991508636870194 1 +192 0 -3.7562728 0.027225705 0.039822987784830025 0 +196 0 5.8725023 0.9853531 6.0932620413355663 1 +198 0 -5.232939 0.00841024 0.012184721289626915 0 +199 0 -4.3234034 0.017386708 0.025304340651562268 0 +201 1 9.770606 0.99936444 0.00091721700916576874 1 +202 0 -4.54813 0.01454009 0.021130913771998817 0 +204 0 -4.54813 0.01454009 0.021130913771998817 0 +205 1 12.087651 0.9999023 0.00014094666717597959 1 +206 1 6.641532 0.99208087 0.01147037136803414 1 +207 0 -4.7285223 0.0125917215 0.018281354087045291 0 +209 0 -3.5942607 0.030918159 0.045309584861749505 0 +210 1 13.547517 0.99997 4.3254285101545484E-05 1 +211 1 9.089206 0.9988979 0.0015908564017088566 1 +212 0 -4.54813 0.01454009 0.021130913771998817 0 +216 0 -4.8905344 0.011062949 0.016049403073119514 0 +218 1 7.7499723 0.99675274 0.0046924303126549046 1 +219 0 -2.4297438 0.07561278 0.11343078610029891 0 +223 1 5.4305964 0.97919416 0.030333134890891962 1 +226 1 9.166205 0.99896437 0.0014948734814852957 1 +228 0 -4.7285223 0.0125917215 0.018281354087045291 0 +233 1 5.6998796 0.9831963 0.02444858173710833 1 +237 1 6.476473 0.99096054 0.01310048672532375 1 +239 1 5.1697493 0.97443426 0.037363242337555155 1 +240 0 -1.9057703 0.11106795 0.16985495245944843 0 +241 0 -3.8436403 0.025416138 0.037141761030701749 0 +242 0 -4.2645645 0.01821836 0.026525908568759592 0 +244 0 -4.54813 0.01454009 0.021130913771998817 0 +246 1 11.068807 0.9997774 0.00032121335660955822 1 +247 1 3.1855059 0.8845626 0.17696383075187846 1 +248 0 -2.7545462 0.059183516 0.088014756501143182 0 249 0 ? ? ? 0 -250 0 -4.652541 0.013439459 0.019520508899098645 0 -252 0 3.7707853 0.9399203 4.0569787298293649 1 -254 1 7.661195 0.9975384 0.0035557339658115286 1 -257 0 -4.334039 0.017470764 0.025427758243753987 0 -258 0 -4.0172215 0.022652287 0.033056170848702819 0 -259 0 4.008581 0.95021814 4.3282360862916729 1 -260 1 7.7233286 0.9976628 0.0033758384253855024 1 -262 1 9.050584 0.9992287 0.0011131570864691799 1 -267 1 3.073165 0.8972062 0.15648852541997652 1 -268 1 8.364619 0.9986318 0.0019752824987132106 1 -269 0 -4.5995197 0.014040394 0.020399552811077585 0 -271 0 -3.4349236 0.036351655 0.053421320856594394 0 -272 1 3.073165 0.8972062 0.15648852541997652 1 +250 0 -4.655863 0.013343408 0.01938005569558102 0 +252 0 4.2659445 0.9483331 4.2746154342457627 1 +254 1 7.919627 0.99716777 0.0040918470836742334 1 +257 0 -4.3234034 0.017386708 0.025304340651562268 0 +258 0 -3.9809995 0.022805596 0.033282492654924653 0 +259 0 4.3407555 0.95121753 4.3574934457139323 1 +260 1 8.330071 0.99796593 0.0029375285520939812 1 +262 1 9.629434 0.99928766 0.0010280485312489948 1 +267 1 3.677929 0.91941935 0.1212050683714937 1 +268 1 8.609313 0.99837637 0.002344307364667635 1 +269 0 -4.54813 0.01454009 0.021130913771998817 0 +271 0 -3.413869 0.0356 0.052296442645105176 0 +272 1 3.677929 0.91941935 0.1212050683714937 1 275 0 ? ? ? 0 -276 0 -4.334039 0.017470764 0.025427758243753987 0 -277 0 -4.916337 0.010806993 0.015676053599848205 0 -278 0 -4.5995197 0.014040394 0.020399552811077585 0 -279 1 6.3361444 0.99257946 0.010745494226668283 1 -280 0 -4.0172215 0.022652287 0.033056170848702819 0 -283 1 5.09361 0.97929937 0.030178143631386233 1 -284 1 5.780157 0.9882371 0.017070901886648178 1 -285 1 12.663693 0.99996245 5.4175552199232265E-05 1 -288 1 1.8098211 0.7520857 0.41103105620854608 1 -290 0 -5.233155 0.008311939 0.012041708383174634 0 -291 0 -4.5995197 0.014040394 0.020399552811077585 0 -293 1 4.209258 0.9575848 0.062527845215913494 1 -296 0 1.7716074 0.74607766 1.977540748016817 1 +276 0 -4.3234034 0.017386708 0.025304340651562268 0 +277 0 -4.8905344 0.011062949 0.016049403073119514 0 +278 0 -4.54813 0.01454009 0.021130913771998817 0 +279 1 6.636298 0.99204755 0.011518825037205057 1 +280 0 -3.9809995 0.022805596 0.033282492654924653 0 +283 1 5.5930414 0.98170805 0.026634049742924003 1 +284 1 6.3683043 0.9901426 0.01429180181912721 1 +285 1 13.342931 0.9999646 5.1079751554952729E-05 1 +288 1 2.133194 0.7659499 0.38467805559449236 1 +290 0 -5.232939 0.00841024 0.012184721289626915 0 +291 0 -4.54813 0.01454009 0.021130913771998817 0 +293 1 4.659646 0.9618815 0.056068896729700848 1 +296 0 2.0631151 0.7556402 2.0329211816413606 1 297 0 ? ? ? 0 -299 1 5.4912786 0.98506975 0.021702211002795221 1 -300 1 6.2749596 0.99219286 0.01130751309089708 1 -301 0 -4.5995197 0.014040394 0.020399552811077585 0 -303 0 -4.5995197 0.014040394 0.020399552811077585 0 -304 1 4.320197 0.961198 0.057094489864426744 1 -308 1 6.5411158 0.9937414 0.0090576348997771245 1 -309 0 -1.7547832 0.13330571 0.2064049009534781 0 -311 0 -5.233155 0.008311939 0.012041708383174634 0 -312 1 3.2156925 0.9076929 0.1397238075599897 1 -314 0 -5.102334 0.009264263 0.01342780259121122 0 -316 1 3.9409618 0.9474733 0.07784282539324304 1 -317 1 8.260409 0.9985074 0.002155003471966795 1 -319 0 2.46204 0.83961296 2.6403705320199373 1 +299 1 6.110527 0.9878861 0.017583423507301729 1 +300 1 6.6979456 0.9924313 0.010960882920314604 1 +301 0 -4.54813 0.01454009 0.021130913771998817 0 +303 0 -4.54813 0.01454009 0.021130913771998817 0 +304 1 4.865142 0.967527 0.047626214248213887 1 +308 1 6.804652 0.99305254 0.010058042410768688 1 +309 0 -1.4760432 0.15027462 0.23493143734082567 0 +311 0 -5.232939 0.00841024 0.012184721289626915 0 +312 1 3.2489738 0.88969976 0.16860953636335851 1 +314 0 -5.070926 0.009575992 0.013881808647455651 0 +316 1 4.409379 0.9537285 0.068349469891991879 1 +317 1 8.798545 0.9986064 0.0020119654595983538 1 +319 0 2.7989092 0.8486183 2.7237375536301185 1 321 0 ? ? ? 0 -323 1 4.269208 0.95957553 0.059531720935566447 1 -327 0 -4.916337 0.010806993 0.015676053599848205 0 -328 1 4.1030073 0.9538244 0.068204404407426977 1 -329 1 6.3562107 0.9927021 0.010567212063419162 1 -331 0 -3.1436715 0.045919728 0.067817441008186471 0 -332 0 -2.8411217 0.058372296 0.086771328388872637 0 -333 1 4.4240284 0.96431077 0.052429940902530503 1 -336 1 4.790291 0.97347915 0.038778012730613723 1 -338 0 -5.102334 0.009264263 0.01342780259121122 0 -343 0 -5.233155 0.008311939 0.012041708383174634 0 -344 1 8.780366 0.99903333 0.0013952818117796901 1 -346 0 -2.788134 0.06085681 0.090582956925044589 0 -347 0 -5.0515347 0.009662537 0.014007879416507383 0 -348 1 -0.033994675 0.3935072 1.3455380164634059 0 -349 1 2.9944906 0.89097595 0.16654160161871584 1 -350 0 -3.776164 0.02757359 0.040339017891068162 0 -352 0 1.3297043 0.6699916 1.5994254020964642 1 -353 1 8.744256 0.9990037 0.0014380614879137648 1 -354 0 -4.916337 0.010806993 0.015676053599848205 0 -355 0 -3.6730852 0.029982215 0.043916895635120724 0 -358 1 6.1854076 0.9915907 0.012183385249969293 1 -360 1 14.4099455 0.9999913 1.2554788140693439E-05 1 -361 1 6.113164 0.99107146 0.012939006126914966 1 -366 1 12.847375 0.99996775 4.6522057140172645E-05 1 -368 0 -4.568268 0.014406925 0.020935975619319223 0 -370 0 -3.0281096 0.0503457 0.074525666475348223 0 -371 0 -4.568268 0.014406925 0.020935975619319223 0 -373 0 -3.605544 0.03166974 0.046428915433668644 0 -376 0 -4.916337 0.010806993 0.015676053599848205 0 -377 0 -5.102334 0.009264263 0.01342780259121122 0 -378 0 -3.5256414 0.033784896 0.049583689746906918 0 -379 0 -1.5692582 0.15227751 0.23833604010598478 0 -381 1 8.122036 0.9983245 0.0024192434292353501 1 -383 0 -4.942006 0.010579844 0.015344805197525327 0 -384 0 -4.942006 0.010579844 0.015344805197525327 0 -387 0 -1.9415016 0.11627045 0.17832317382513535 0 -388 0 -4.44138 0.01599441 0.023261582227407044 0 -389 0 -2.9747353 0.05252372 0.07783827045445077 0 -391 1 8.779809 0.99903286 0.0013959704081915206 1 -392 0 -4.334039 0.017470764 0.025427758243753987 0 -395 0 -4.334039 0.017470764 0.025427758243753987 0 -396 0 -4.0172215 0.022652287 0.033056170848702819 0 -398 0 -3.902061 0.024885992 0.036357189097323071 0 -399 0 -4.3200974 0.017672086 0.025723399534110371 0 -404 0 -4.508781 0.015130844 0.021996025629370697 0 -406 0 -3.4622765 0.035558585 0.052234490586920097 0 -409 0 -4.0015955 0.022943486 0.033486082550290931 0 -413 0 -3.1024804 0.04745315 0.070138043352784826 0 -414 1 5.959919 0.9898627 0.014699694943656067 1 -415 0 -0.721817 0.26737925 0.44886153992419614 0 -416 1 8.443301 0.99871886 0.0018494825270406309 1 -418 0 -1.8258505 0.12658581 0.19526212698148004 0 -419 0 -3.8876748 0.025179695 0.03679179342594642 0 -422 0 -2.1972284 0.09602885 0.14565136391673347 0 -423 0 -3.0868545 0.04804749 0.071038491836753939 0 -428 0 -4.916337 0.010806993 0.015676053599848205 0 -429 0 -4.6251884 0.013746234 0.019969190332632794 0 -430 0 -4.2361884 0.018932736 0.027576040231473427 0 -434 0 5.330062 0.982951 5.874167835772826 1 -436 1 4.9601755 0.9769113 0.03370050879649044 1 -439 0 -4.068559 0.021720776 0.031681792181227325 0 -440 1 7.0005217 0.9957298 0.0061737815665520758 1 -441 0 -1.8277493 0.12641032 0.19497228221884577 0 -442 0 -4.126358 0.02071682 0.030201989467604867 0 -449 1 9.384189 0.9994164 0.00084218683672760905 1 -450 0 -3.936492 0.024196552 0.03533751402030335 0 -451 0 -4.068559 0.021720776 0.031681792181227325 0 -452 0 -4.358462 0.017123515 0.024917965570564662 0 -453 1 7.3491344 0.99680644 0.0046147018108436385 1 -454 0 -4.259669 0.018571336 0.027044686673549208 0 -455 1 0.2950573 0.46074596 1.1179565776036025 1 -456 1 8.340758 0.99860424 0.0020150654708657578 1 -457 1 7.996641 0.99813956 0.0026865470958503841 1 -464 0 -4.3597083 0.017105974 0.024892219294896437 0 -465 1 8.680116 0.9989488 0.001517340659881941 1 -466 1 8.110646 0.9983085 0.0024424141036858663 1 -467 1 6.858451 0.9951935 0.0069510590946110862 1 -474 0 -4.068559 0.021720776 0.031681792181227325 0 -480 0 -4.254556 0.018649446 0.02715951257322859 0 -482 1 13.881022 0.9999864 1.9606155421105871E-05 1 -483 1 9.617421 0.9995198 0.00069291249062615099 1 -484 0 -3.736116 0.028486205 0.041693613791526342 0 -487 1 11.720016 0.99991727 0.00011936089720480338 1 -489 1 -0.5905738 0.28942576 1.7887347593944831 0 -492 0 -4.228887 0.019046513 0.027743364345069858 0 -493 1 9.492114 0.9994668 0.0007694835687733193 1 -495 0 -4.520036 0.01499118 0.021791452089244788 0 -497 0 -4.111538 0.020969829 0.030574773902338735 0 -501 0 -4.0428905 0.02218175 0.032361761861453661 0 -502 0 -3.8966928 0.024995195 0.036518765772720306 0 -504 0 -5.233155 0.008311939 0.012041708383174634 0 -507 0 -4.0115523 0.022757515 0.033211510122296124 0 -510 0 -5.233155 0.008311939 0.012041708383174634 0 -513 0 -4.520036 0.01499118 0.021791452089244788 0 -514 1 8.787938 0.9990394 0.001386502236338372 1 -517 0 -5.102334 0.009264263 0.01342780259121122 0 -519 1 6.320156 0.99248034 0.010889574092143747 1 -520 0 -5.0471582 0.009697631 0.014059004390630944 0 -521 0 -4.1737113 0.019928277 0.029040762705050325 0 -522 1 3.983387 0.94921166 0.075198276844941334 1 -523 1 6.156104 0.9913838 0.012484423812211912 1 -527 0 -3.7260728 0.028719628 0.042040287835734616 0 -528 0 -2.9663248 0.05287494 0.078373162815022712 0 -529 0 -4.228887 0.019046513 0.027743364345069858 0 -531 0 -3.4622765 0.035558585 0.052234490586920097 0 -532 0 -4.7855167 0.012041707 0.01747795506449706 0 -533 0 -4.334039 0.017470764 0.025427758243753987 0 -534 0 -4.6251884 0.013746234 0.019969190332632794 0 -535 0 -3.7884345 0.0272997 0.039932732580676075 0 -538 0 -4.0428905 0.02218175 0.032361761861453661 0 -539 0 -3.4605927 0.035606917 0.052306791467165305 0 -540 0 -3.3445406 0.039095003 0.057534293815457598 0 -541 0 -4.650857 0.013458151 0.019547844212221475 0 -544 0 -3.8141036 0.026735267 0.039095815560991697 0 -546 1 10.355874 0.999741 0.00037368069940970906 1 -547 0 -5.128003 0.009069243 0.013143845413371364 0 -548 0 -4.836854 0.011541378 0.016747520160516439 0 -549 1 5.2726183 0.9821267 0.026018923571932827 1 -557 0 -3.776164 0.02757359 0.040339017891068162 0 -558 0 -4.6251884 0.013746234 0.019969190332632794 0 -559 0 -3.7517414 0.028126696 0.041159842562845529 0 -560 0 -3.4349236 0.036351655 0.053421320856594394 0 -561 0 -3.4349236 0.036351655 0.053421320856594394 0 -563 0 -4.334039 0.017470764 0.025427758243753987 0 -565 1 10.456666 0.999762 0.00034340422940330219 1 -566 0 -3.6847782 0.029699048 0.043495806785783089 0 -569 1 8.855367 0.99909204 0.0013105008899586513 1 -577 0 -4.916337 0.010806993 0.015676053599848205 0 -578 0 -4.916337 0.010806993 0.015676053599848205 0 -581 1 8.00238 0.99814844 0.0026737105637673411 1 -582 1 7.645852 0.9975067 0.0036015949700722869 1 -584 0 -3.1515741 0.04563098 0.067380883189205654 0 -586 1 12.260621 0.99994737 7.5932338567167971E-05 1 -590 1 4.0090714 0.9502375 0.073639933647810851 1 -593 0 -3.736116 0.028486205 0.041693613791526342 0 -594 1 5.269803 0.98208535 0.02607968888447142 1 -600 0 -4.334039 0.017470764 0.025427758243753987 0 -602 0 -4.0428905 0.02218175 0.032361761861453661 0 -604 1 4.515382 0.9668494 0.048636927001133007 1 -606 0 -4.2135105 0.019288328 0.028099045966060599 0 -607 0 -5.233155 0.008311939 0.012041708383174634 0 -609 0 -4.068559 0.021720776 0.031681792181227325 0 -612 1 14.881892 0.9999941 8.5131663352343685E-06 1 -613 0 -4.128848 0.020674605 0.030139799036072543 0 -614 0 -4.8111854 0.01178892 0.017108862624922505 0 +323 1 4.8855343 0.968041 0.046859938697557786 1 +327 0 -4.8905344 0.011062949 0.016049403073119514 0 +328 1 4.1213064 0.94229364 0.08575138254712171 1 +329 1 6.918497 0.99365956 0.0091764494650447715 1 +331 0 -3.125589 0.044527918 0.065714375419183207 0 +332 0 -2.8317442 0.055802386 0.082839258038727215 0 +333 1 4.9879713 0.97050625 0.043190589776101178 1 +336 1 5.3119774 0.9771479 0.033351185116665787 1 +338 0 -5.070926 0.009575992 0.013881808647455651 0 +343 0 -5.232939 0.00841024 0.012184721289626915 0 +344 1 9.373975 0.99912435 0.0012638519898626412 1 +346 0 -2.8051786 0.056944884 0.084586005071702747 0 +347 0 -5.0361757 0.009846162 0.014275404034477916 0 +348 1 0.09843922 0.38710535 1.3692018642420059 1 +349 1 3.543579 0.9109965 0.13448258886866615 1 +350 0 -3.7809267 0.02670275 0.039047616782150546 0 +352 0 1.5155859 0.66513264 1.5783383412583629 1 +353 1 9.391765 0.99913687 0.001245778098200614 1 +354 0 -4.8905344 0.011062949 0.016049403073119514 0 +355 0 -3.6708684 0.029115355 0.042628201360876333 0 +358 1 6.2604885 0.9892544 0.015586495937560322 1 +360 1 15.035466 0.999991 1.2984748573554836E-05 1 +361 1 6.5571547 0.99152654 0.012276699619017499 1 +366 1 13.467439 0.999968 4.6178080787551932E-05 1 +368 0 -4.532379 0.01472369 0.021399725812029809 0 +370 0 -3.0555735 0.04699932 0.069450851924502996 0 +371 0 -4.532379 0.01472369 0.021399725812029809 0 +373 0 -3.5973978 0.030842254 0.04519658823690309 0 +376 0 -4.8905344 0.011062949 0.016049403073119514 0 +377 0 -5.070926 0.009575992 0.013881808647455651 0 +378 0 -3.4778075 0.033867203 0.04970659025721904 0 +379 0 -1.5486526 0.14293145 0.22251749039141275 0 +381 1 8.643491 0.99842054 0.0022804855789229006 1 +383 0 -4.9493732 0.010554424 0.015307739563326032 0 +384 0 -4.9493732 0.010554424 0.015307739563326032 0 +387 0 -1.9204488 0.10990166 0.16796335660581804 0 +388 0 -4.418391 0.016122172 0.023448912843152432 0 +389 0 -2.9322662 0.051671732 0.076541554407895826 0 +391 1 9.323798 0.9990881 0.001316181486351004 1 +392 0 -4.3234034 0.017386708 0.025304340651562268 0 +395 0 -4.3234034 0.017386708 0.025304340651562268 0 +396 0 -3.9809995 0.022805596 0.033282492654924653 0 +398 0 -3.9080298 0.024158016 0.03528054047560631 0 +399 0 -4.305583 0.017634572 0.025668306162224368 0 +404 0 -4.4998136 0.015110562 0.021966315150339356 0 +406 0 -3.4627619 0.03426748 0.050304437450714536 0 +409 0 -3.973124 0.022947924 0.033492636604193209 0 +413 0 -3.063589 0.046709906 0.069012790465941026 0 +414 1 6.5001116 0.9911302 0.01285354412946392 1 +415 0 -0.5379734 0.27407813 0.46211381508363975 0 +416 1 8.981729 0.99879795 0.0017352301830455481 1 +418 0 -1.7460232 0.124474354 0.1917786561625977 0 +419 0 -3.9546041 0.02328605 0.033991992887674995 0 +422 0 -2.1661267 0.09192332 0.13911396597120457 0 +423 0 -3.0557137 0.046994247 0.069443170930934078 0 +428 0 -4.8905344 0.011062949 0.016049403073119514 0 +429 0 -4.606969 0.013873975 0.020156062560767365 0 +430 0 -4.248364 0.018454121 0.026872392407670757 0 +434 0 5.678664 0.9829106 5.8707522055318933 1 +436 1 4.8968496 0.9683228 0.046440010390598266 1 +439 0 -4.098677 0.02077887 0.030293405816888826 0 +440 1 7.6677313 0.99653023 0.0050145176664260493 1 +441 0 -1.6798639 0.13042165 0.20161207865646949 0 +442 0 -4.0051136 0.022375137 0.032647118008373256 0 +449 1 10.087396 0.99950796 0.00071003308680625738 1 +450 0 -3.9024792 0.024264034 0.035437287169513824 0 +451 0 -4.098677 0.02077887 0.030293405816888826 0 +452 0 -4.348057 0.017049437 0.024809236682836164 0 +453 1 8.068194 0.9974875 0.0036293536545799203 1 +454 0 -4.315388 0.017497765 0.025467405517597397 0 +455 1 0.4596901 0.45824033 1.1258236575305995 1 +456 1 8.942605 0.9987594 0.0017909346040298939 1 +457 1 8.346686 0.99799305 0.0028983232844613941 1 +464 0 -4.382242 0.01659238 0.024138560366440985 0 +465 1 8.953591 0.99877036 0.0017750926332974581 1 +466 1 8.690221 0.998479 0.0021959971123143552 1 +467 1 7.4474783 0.99585676 0.0059898462794332257 1 +474 0 -4.098677 0.02077887 0.030293405816888826 0 +480 0 -4.279069 0.018009793 0.026219457722364156 0 +482 1 14.295785 0.9999836 2.3647808303057529E-05 1 +483 1 10.273198 0.99957657 0.00061101174577425621 1 +484 0 -3.7483978 0.027394837 0.040073844510940371 0 +487 1 12.309564 0.99991834 0.00011781292611185561 1 +489 1 -0.43852377 0.29036227 1.7840741013188408 0 +492 0 -4.22023 0.018870676 0.027484782205950953 0 +493 1 10.096398 0.99951154 0.00070487107661414573 1 +495 0 -4.5037956 0.015062724 0.02189624324376066 0 +497 0 -4.155446 0.019865254 0.028947994527876966 0 +501 0 -4.039838 0.021769192 0.03175319395295028 0 +502 0 -3.8809624 0.024679318 0.036051445088729372 0 +504 0 -5.232939 0.00841024 0.012184721289626915 0 +507 0 -3.8678274 0.024936227 0.03643151520715443 0 +510 0 -5.232939 0.00841024 0.012184721289626915 0 +513 0 -4.5037956 0.015062724 0.02189624324376066 0 +514 1 9.508458 0.99921453 0.0011336389645461593 1 +517 0 -5.070926 0.009575992 0.013881808647455651 0 +519 1 6.943595 0.99378616 0.0089926501548524794 1 +520 0 -5.0525465 0.009717959 0.014088618955572644 0 +521 0 -4.2018504 0.019147774 0.027892297047422871 0 +522 1 4.3094034 0.95002776 0.073958419155255584 1 +523 1 6.7598047 0.99279785 0.010428101441132083 1 +527 0 -3.697434 0.028514326 0.041735373036084926 0 +528 0 -2.955677 0.05075209 0.07514317509343231 0 +529 0 -4.22023 0.018870676 0.027484782205950953 0 +531 0 -3.4627619 0.03426748 0.050304437450714536 0 +532 0 -4.7285223 0.0125917215 0.018281354087045291 0 +533 0 -4.3234034 0.017386708 0.025304340651562268 0 +534 0 -4.606969 0.013873975 0.020156062560767365 0 +535 0 -3.6886954 0.028710688 0.042027007798555546 0 +538 0 -4.039838 0.021769192 0.03175319395295028 0 +539 0 -3.4727077 0.03400237 0.049908447053141368 0 +540 0 -3.3012362 0.03886211 0.057184671042155688 0 +541 0 -4.6658077 0.013237966 0.019225886685070071 0 +544 0 -3.7475338 0.027413454 0.040101460146895203 0 +546 1 11.016191 0.9997677 0.00033514712058262439 1 +547 0 -5.1297655 0.009135182 0.013239848501308635 0 +548 0 -4.8462 0.011462052 0.016631746199728435 0 +549 1 5.5535192 0.9811253 0.027490706489269167 1 +557 0 -3.7809267 0.02670275 0.039047616782150546 0 +558 0 -4.606969 0.013873975 0.020156062560767365 0 +559 0 -3.7562728 0.027225705 0.039822987784830025 0 +560 0 -3.413869 0.0356 0.052296442645105176 0 +561 0 -3.413869 0.0356 0.052296442645105176 0 +563 0 -4.3234034 0.017386708 0.025304340651562268 0 +565 1 11.259841 0.99980927 0.00027519848718779308 1 +566 0 -3.6307197 0.030047078 0.04401336861004132 0 +569 1 9.242663 0.99902636 0.0014053525670487904 1 +577 0 -4.8905344 0.011062949 0.016049403073119514 0 +578 0 -4.8905344 0.011062949 0.016049403073119514 0 +581 1 8.654239 0.9984342 0.0022607625481964131 1 +582 1 8.128011 0.99760586 0.0034581548754549166 1 +584 0 -3.123908 0.04458577 0.065801733133594345 0 +586 1 12.971318 0.9999522 6.8966691413912035E-05 1 +590 1 4.1888075 0.94519055 0.081322891229328675 1 +593 0 -3.7483978 0.027394837 0.040073844510940371 0 +594 1 5.7850237 0.98429656 0.022835041928152281 1 +600 0 -4.3234034 0.017386708 0.025304340651562268 0 +602 0 -4.039838 0.021769192 0.03175319395295028 0 +604 1 5.138941 0.9738064 0.038293140457828272 1 +606 0 -4.2233667 0.01882378 0.027415826214124143 0 +607 0 -5.232939 0.00841024 0.012184721289626915 0 +609 0 -4.098677 0.02077887 0.030293405816888826 0 +612 1 15.565835 0.99999416 8.427174504939984E-06 1 +613 0 -4.153376 0.019897865 0.028995996893701973 0 +614 0 -4.787361 0.012013777 0.017437170868677548 0 617 0 ? ? ? 0 -618 0 -4.0428905 0.02218175 0.032361761861453661 0 -619 0 -3.7517414 0.028126696 0.041159842562845529 0 -621 0 -0.14650154 0.3712855 0.66952305434617909 0 -622 0 -2.5446053 0.07359522 0.11028539506852417 0 -624 0 -3.7915406 0.02723079 0.039830529260958501 0 -627 0 -3.3132024 0.04009178 0.059031621926170716 0 -629 0 -4.3597083 0.017105974 0.024892219294896437 0 -633 1 4.0237722 0.9508158 0.072762222474825514 1 -634 0 -4.650857 0.013458151 0.019547844212221475 0 -638 0 -4.3597083 0.017105974 0.024892219294896437 0 -639 0 -3.776164 0.02757359 0.040339017891068162 0 -641 0 -4.334039 0.017470764 0.025427758243753987 0 -642 0 -4.334039 0.017470764 0.025427758243753987 0 -644 0 -4.942006 0.010579844 0.015344805197525327 0 -645 0 -4.334039 0.017470764 0.025427758243753987 0 -649 0 -4.334039 0.017470764 0.025427758243753987 0 -652 0 -3.567933 0.032648917 0.047888509876543181 0 -653 0 -4.0428905 0.02218175 0.032361761861453661 0 -654 0 -4.0172215 0.022652287 0.033056170848702819 0 -656 0 -3.7517414 0.028126696 0.041159842562845529 0 -657 0 0.674386 0.5399075 1.1200042029157564 1 -660 0 -4.916337 0.010806993 0.015676053599848205 0 -661 0 -3.7260728 0.028719628 0.042040287835734616 0 -665 0 -5.233155 0.008311939 0.012041708383174634 0 -668 1 3.2994661 0.91339904 0.13068282036653511 1 -670 1 6.4614477 0.99331295 0.0096797675127115335 1 -678 0 -5.233155 0.008311939 0.012041708383174634 0 -679 0 -4.942006 0.010579844 0.015344805197525327 0 -680 1 14.404435 0.99999124 1.2640780217014572E-05 1 -681 1 9.278363 0.99936247 0.00092005653039600916 1 -682 0 -3.2511153 0.04213895 0.06211170367011902 0 -683 0 -5.233155 0.008311939 0.012041708383174634 0 -685 0 -5.233155 0.008311939 0.012041708383174634 0 -688 0 -4.3597083 0.017105974 0.024892219294896437 0 -689 0 -3.1943884 0.044096444 0.065063026905048796 0 -691 1 5.2444315 0.9817081 0.026633962149343771 1 -692 0 -4.650857 0.013458151 0.019547844212221475 0 -693 0 -4.042787 0.022183627 0.032364532038170898 0 -694 0 -4.057659 0.02191537 0.031968794192829511 0 -696 1 7.3569994 0.99682736 0.004584422473559115 1 -697 1 4.656295 0.9704265 0.043309147612224416 1 -698 1 5.6929607 0.9873583 0.018354419974050573 1 -0 0 -3.4721131 0.041149348 0.060621972477918451 0 -1 0 2.4163914 0.8102447 2.3977877569564514 1 -2 0 -4.045404 0.026690751 0.039029830826214376 0 -3 0 2.9251795 0.86401767 2.8785088509147978 1 +618 0 -4.039838 0.021769192 0.03175319395295028 0 +619 0 -3.7562728 0.027225705 0.039822987784830025 0 +621 0 0.0042414665 0.36919913 0.66474343803847269 1 +622 0 -2.5211115 0.070608765 0.10564205571811729 0 +624 0 -3.7777896 0.026768748 0.039145446220591275 0 +627 0 -3.1292257 0.044402987 0.065525750299537014 0 +629 0 -4.382242 0.01659238 0.024138560366440985 0 +633 1 4.158328 0.9438998 0.083294361183546328 1 +634 0 -4.6658077 0.013237966 0.019225886685070071 0 +638 0 -4.382242 0.01659238 0.024138560366440985 0 +639 0 -3.7809267 0.02670275 0.039047616782150546 0 +641 0 -4.3234034 0.017386708 0.025304340651562268 0 +642 0 -4.3234034 0.017386708 0.025304340651562268 0 +644 0 -4.9493732 0.010554424 0.015307739563326032 0 +645 0 -4.3234034 0.017386708 0.025304340651562268 0 +649 0 -4.3234034 0.017386708 0.025304340651562268 0 +652 0 -3.5676956 0.0315682 0.046277640684522559 0 +653 0 -4.039838 0.021769192 0.03175319395295028 0 +654 0 -3.9809995 0.022805596 0.033282492654924653 0 +656 0 -3.7562728 0.027225705 0.039822987784830025 0 +657 0 0.6928787 0.50527525 1.0153020171868359 1 +660 0 -4.8905344 0.011062949 0.016049403073119514 0 +661 0 -3.697434 0.028514326 0.041735373036084926 0 +665 0 -5.232939 0.00841024 0.012184721289626915 0 +668 1 3.362691 0.8984041 0.15456354881831241 1 +670 1 6.9357576 0.9937469 0.0090496738948646453 1 +678 0 -5.232939 0.00841024 0.012184721289626915 0 +679 0 -4.9493732 0.010554424 0.015307739563326032 0 +680 1 15.087735 0.99999136 1.2468796069497877E-05 1 +681 1 9.870067 0.9994135 0.00084640287826628446 1 +682 0 -3.2252913 0.041221604 0.060730693175907312 0 +683 0 -5.232939 0.00841024 0.012184721289626915 0 +685 0 -5.232939 0.00841024 0.012184721289626915 0 +688 0 -4.382242 0.01659238 0.024138560366440985 0 +689 0 -3.214981 0.041552313 0.061228404039596784 0 +691 1 5.504386 0.98037547 0.028593709808496726 1 +692 0 -4.6658077 0.013237966 0.019225886685070071 0 +693 0 -4.035124 0.021850502 0.031873115017123851 0 +694 0 -4.030379 0.021932645 0.031994274177204983 0 +696 1 7.6070995 0.99635655 0.0052659907809486433 1 +697 1 5.0033865 0.9708609 0.042663488663692947 1 +698 1 6.1210947 0.98798794 0.017434669917962976 1 +0 0 -3.4721127 0.041149363 0.060621994898335516 0 +1 0 2.4163914 0.8102446 2.3977873037870139 1 +2 0 -4.0454035 0.02669076 0.03902984463081418 0 +3 0 2.9251804 0.8640177 2.8785094832862348 1 4 0 -3.5088277 0.04003246 0.058942472936424177 0 -7 0 -4.670553 0.016548801 0.024074630604726591 0 -12 1 -0.34343147 0.3308405 1.5957922491655303 0 -13 0 -4.6186943 0.017221216 0.025061381581356009 0 +7 0 -4.6705537 0.016548796 0.024074622407383658 0 +12 1 -0.34343147 0.33084044 1.5957925090833267 0 +13 0 -4.618695 0.01722121 0.025061373378404472 0 14 1 7.360214 0.9951006 0.0070856869219221227 1 -15 1 0.6494303 0.517803 0.94952473394442127 1 -16 0 -4.220706 0.023354558 0.03409318909369452 0 -17 0 -3.9551725 0.028584337 0.04183934627600109 0 -19 0 -2.9890537 0.05890274 0.087584266111585152 0 +15 1 0.6494312 0.5178032 0.9495242357357877 1 +16 0 -4.220706 0.023354556 0.034093186342205803 0 +17 0 -3.955172 0.028584346 0.041839360107510271 0 +19 0 -2.9890532 0.058902755 0.087584288954953682 0 22 0 -4.7037654 0.016131794 0.023463022615813851 0 23 1 ? ? ? 0 24 0 -5.4043503 0.009396286 0.013620064546924614 0 26 0 -4.390918 0.020506337 0.029891936543840706 0 27 0 -3.7376466 0.033700433 0.049457579363906538 0 29 0 -5.4339433 0.009183524 0.013310236202903244 0 -30 0 -4.649441 0.016819378 0.024471614226605987 0 -33 0 -4.698118 0.016201962 0.023565916692469265 0 +30 0 -4.649441 0.016819376 0.024471611493406355 0 +33 0 -4.6981187 0.016201956 0.023565908498016312 0 34 0 -4.4530277 0.0195541 0.028490070077080931 0 36 1 7.832773 0.9966079 0.004902085221172487 1 38 1 4.92861 0.96814406 0.046706359368045791 1 -39 1 1.079258 0.60037446 0.73606548851031739 1 +39 1 1.079258 0.6003744 0.73606563173981054 1 42 1 6.8985863 0.9929881 0.010151652400815765 1 -43 1 -0.49528694 0.30512375 1.7125336322001359 0 -47 0 -5.669884 0.0076494967 0.011078316903702529 0 -49 1 5.3024044 0.9760178 0.035020675578791655 1 +43 1 -0.49528742 0.30512363 1.7125341958491029 0 +47 0 -5.669884 0.007649497 0.011078317580688359 0 +49 1 5.3024054 0.9760178 0.035020675578791655 1 53 1 5.116103 0.9723645 0.040430893694146056 1 55 1 4.4195347 0.95331246 0.068978947156115675 1 -57 1 0.5701313 0.5023233 0.99331181510286393 1 -58 1 1.1371031 0.61116564 0.71036465171693308 1 +57 1 0.57013035 0.50232315 0.99331232866456232 1 +58 1 1.137104 0.61116576 0.71036437031591504 1 59 1 1.6442327 0.7002286 0.51410216467535697 1 61 0 -5.2770567 0.010368526 0.015036710692972488 0 -62 1 5.7670774 0.98319566 0.024449543808342401 1 -65 1 2.7867746 0.8508072 0.23309586971174173 1 +62 1 5.7670784 0.9831957 0.024449456347294611 1 +65 1 2.7867756 0.85080725 0.23309576864145995 1 67 1 2.9058514 0.8622339 0.21384884774751767 1 -75 0 -4.291042 0.022133557 0.032290659731254154 0 +75 0 -4.2910423 0.022133548 0.032290645990988666 0 78 0 -3.652576 0.035932966 0.052794630104546485 0 -80 0 -3.2955709 0.046948776 0.069374337067975425 0 +80 0 -3.2955709 0.046948772 0.069374331428763494 0 81 0 -3.9975338 0.027679635 0.040496356631681558 0 -83 0 -2.9832687 0.059153758 0.087969125401985163 0 -84 1 6.5824003 0.99104106 0.012983257475290413 1 +83 0 -2.9832683 0.059153773 0.087969148251448304 0 +84 1 6.5823994 0.99104106 0.012983257475290413 1 85 1 4.7604074 0.9638325 0.053145649556241237 1 -86 1 1.466999 0.6703859 0.5769362953607623 1 -87 1 5.223543 0.97453266 0.037217553175852479 1 -89 0 -5.017977 0.012665044 0.01838848844354141 0 -94 0 -4.9212914 0.013645148 0.019821329360883679 0 -101 1 -0.85990286 0.24827081 2.0100134477430442 0 -103 1 0.36755466 0.4628286 1.1114500580380418 1 +86 1 1.466999 0.67038584 0.57693642363214981 1 +87 1 5.223544 0.97453266 0.037217553175852479 1 +89 0 -5.017977 0.012665043 0.018388487082691734 0 +94 0 -4.921291 0.013645153 0.019821336171893227 0 +101 1 -0.85990334 0.2482707 2.0100140538749183 0 +103 1 0.36755466 0.46282858 1.1114501509356363 1 107 1 4.617798 0.9597418 0.059281809358577985 1 -110 0 -3.1312823 0.053037085 0.078620166419129056 0 -114 0 -2.8847933 0.06358328 0.094777396090560642 0 -116 0 -0.40737772 0.3198761 0.55613051468995855 0 +110 0 -3.1312823 0.05303708 0.078620160743660947 0 +114 0 -2.8847933 0.06358327 0.094777384611786961 0 +116 0 -0.40737724 0.31987616 0.5561306411247593 0 118 0 -5.2584124 0.010519049 0.015256161591804486 0 -119 0 -3.786881 0.03247014 0.047621909253386469 0 -124 1 5.0384197 0.97068614 0.042923205850192866 1 +119 0 -3.7868814 0.03247013 0.047621892588915325 0 +124 1 5.0384197 0.9706861 0.042923294438382942 1 126 1 6.4765244 0.9902761 0.014097276720975665 1 127 0 -4.4382315 0.01977694 0.028818008476875413 0 -130 0 -3.2361722 0.049069233 0.072587786489277134 0 -134 0 -4.75809 0.015471877 0.022495677067487178 0 -135 0 -2.670867 0.07428964 0.1113672306008736 0 -136 0 -4.220706 0.023354558 0.03409318909369452 0 +130 0 -3.2361717 0.04906925 0.072587809096424025 0 +134 0 -4.75809 0.015471876 0.022495675702757793 0 +135 0 -2.670867 0.074289635 0.11136721898934138 0 +136 0 -4.220706 0.023354556 0.034093186342205803 0 139 0 ? ? ? 0 140 0 -4.9692993 0.013149481 0.019096523087584735 0 -142 1 3.7043686 0.92112815 0.11852620676711542 1 -143 0 -4.364107 0.02093128 0.030517968916693023 0 -146 1 0.36148834 0.4616506 1.1151267020337663 1 -148 0 -2.357305 0.092992336 0.14081335294741762 0 -149 1 8.756336 0.9983485 0.0023846173156615046 1 +142 1 3.7043686 0.9211281 0.11852630012147426 1 +143 0 -4.364107 0.020931277 0.030517966172014473 0 +146 1 0.3614874 0.4616504 1.1151273539763837 1 +148 0 -2.3573046 0.09299236 0.14081338850031172 0 +149 1 8.756334 0.9983485 0.0023846173156615046 1 153 0 -3.7005844 0.03465612 0.050885134255065932 0 155 1 3.1619148 0.88432163 0.17735691801298636 1 -157 0 -4.9212914 0.013645148 0.019821329360883679 0 +157 0 -4.921291 0.013645153 0.019821336171893227 0 158 0 ? ? ? 0 159 1 10.452139 0.99956036 0.00063441148415971371 1 160 1 7.997595 0.9970165 0.0043107295502774808 1 162 0 -4.4382315 0.01977694 0.028818008476875413 0 163 0 -3.669045 0.035489928 0.052131791002515318 0 -165 0 -3.3581352 0.04480965 0.066139834275042608 0 +165 0 -3.3581352 0.044809647 0.066139828648459548 0 166 1 6.3355865 0.9891566 0.015729147554240097 1 168 0 -4.4382315 0.01977694 0.028818008476875413 0 170 0 -4.9692993 0.013149481 0.019096523087584735 0 -172 0 -5.669884 0.0076494967 0.011078316903702529 0 -175 1 6.0381765 0.98635924 0.019814913404171304 1 -178 0 -3.9551725 0.028584337 0.04183934627600109 0 -182 0 -2.9890537 0.05890274 0.087584266111585152 0 +172 0 -5.669884 0.007649497 0.011078317580688359 0 +175 1 6.0381775 0.98635924 0.019814913404171304 1 +178 0 -3.955172 0.028584346 0.041839360107510271 0 +182 0 -2.9890532 0.058902755 0.087584288954953682 0 184 1 5.409273 0.9778962 0.032246738266973496 1 185 0 -4.8669662 0.014228282 0.020674503875052902 0 186 1 3.9876003 0.93577826 0.095761382134015888 1 190 1 10.521242 0.9995835 0.00060103256102277266 1 193 0 -5.4043503 0.009396286 0.013620064546924614 0 194 0 -4.4382315 0.01977694 0.028818008476875413 0 -195 0 -3.9551725 0.028584337 0.04183934627600109 0 +195 0 -3.955172 0.028584346 0.041839360107510271 0 197 0 -2.8025956 0.067515396 0.10084818862163951 0 200 1 8.673521 0.99823827 0.002543887091121199 1 -203 0 -3.4721131 0.041149348 0.060621972477918451 0 -208 0 -5.350025 0.009799645 0.014207628773175909 0 +203 0 -3.4721127 0.041149363 0.060621994898335516 0 +208 0 -5.3500257 0.009799642 0.014207623345529065 0 213 1 12.300528 0.9998962 0.00014971866598967564 1 214 1 11.895983 0.99985766 0.00020536190066155745 1 -215 1 6.600219 0.99116385 0.012804525065380799 1 +215 1 6.6002197 0.99116385 0.012804525065380799 1 217 0 -5.4043503 0.009396286 0.013620064546924614 0 220 0 -5.181178 0.011165957 0.016199682480376523 0 221 1 7.9662914 0.9969428 0.0044173368194155311 1 -222 1 -2.1487255 0.10767716 3.2152158079150412 0 -224 1 8.4735565 0.997941 0.0029735466380567798 1 -225 0 -5.669884 0.0076494967 0.011078316903702529 0 +222 1 -2.1487255 0.107677154 3.2152159077404439 0 +224 1 8.473555 0.997941 0.0029735466380567798 1 +225 0 -5.669884 0.007649497 0.011078317580688359 0 227 1 6.748211 0.992121 0.011412038429712602 1 229 1 10.504805 0.9995781 0.00060877502594878564 1 -230 1 4.829337 0.9656635 0.050407558021295686 1 +230 1 4.829338 0.9656635 0.050407558021295686 1 231 1 6.912092 0.9930612 0.010045486491117618 1 232 0 1.0722923 0.59906816 1.3185711185935187 1 -234 0 -2.7037287 0.07254336 0.10864826020710561 0 +234 0 -2.7037287 0.07254335 0.10864824861743642 0 235 0 ? ? ? 0 236 1 9.440506 0.9990316 0.0013977779753363091 1 238 1 10.690645 0.9996351 0.00052653496634314379 1 -243 0 -3.301972 0.046725534 0.069036441226174985 0 -245 0 -2.856113 0.06493041 0.096854356008988451 0 +243 0 -3.3019714 0.04672555 0.069036463777740448 0 +245 0 -2.8561125 0.064930424 0.096854378999610469 0 251 1 7.355525 0.99508274 0.0071116115660869242 1 253 1 6.8985863 0.9929881 0.010151652400815765 1 -255 1 3.745204 0.9234148 0.11494919833274581 1 +255 1 3.745205 0.9234149 0.11494910520956844 1 256 0 -4.9692993 0.013149481 0.019096523087584735 0 261 1 9.009869 0.9986448 0.0019564247198453513 1 263 1 7.1455708 0.9942114 0.0083754842729617477 1 264 1 4.0145664 0.9370327 0.093828700465107676 1 -265 0 -2.5156913 0.0830686 0.12511429472057489 0 +265 0 -2.5156918 0.083068565 0.12511423610705438 0 266 1 7.325534 0.99496675 0.0072797874039680641 1 270 1 5.5723915 0.9804892 0.028426363787945077 1 -273 1 0.03731823 0.39964545 1.3232074389645836 1 -274 0 -4.2340226 0.023118448 0.033744449784985975 0 -281 0 -4.698118 0.016201962 0.023565916692469265 0 -282 1 2.860156 0.8579384 0.2210540141016637 1 +273 1 0.037317276 0.39964524 1.3232081920564036 1 +274 0 -4.2340226 0.023118446 0.033744447034162281 0 +281 0 -4.6981187 0.016201956 0.023565908498016312 0 +282 1 2.860157 0.85793847 0.22105391387148221 1 286 1 12.544172 0.9999142 0.00012374682432127928 1 -287 0 -4.75809 0.015471877 0.022495677067487178 0 -289 1 6.6595697 0.9915608 0.012226832918130892 1 +287 0 -4.75809 0.015471876 0.022495675702757793 0 +289 1 6.6595707 0.9915608 0.012226832918130892 1 292 1 ? ? ? 0 294 0 ? ? ? 0 295 1 5.621522 0.98121005 0.027366079814265764 1 -298 0 -2.4584546 0.086538464 0.13058411343083171 0 -302 1 12.725584 0.99992555 0.0001074071635410925 1 +298 0 -2.4584541 0.086538486 0.13058414873253427 0 +302 1 12.72558 0.99992555 0.0001074071635410925 1 305 1 8.040865 0.9971154 0.0041676501364772073 1 306 0 -5.4043503 0.009396286 0.013620064546924614 0 307 0 -5.4043503 0.009396286 0.013620064546924614 0 @@ -511,35 +511,35 @@ Instance Label Score Probability Log-loss Assigned 320 1 5.5611877 0.98032105 0.028673793676042177 1 322 0 -4.4382315 0.01977694 0.028818008476875413 0 324 0 -5.4043503 0.009396286 0.013620064546924614 0 -325 0 -3.7860875 0.03248962 0.047650955719163814 0 -326 1 3.6223288 0.91634446 0.12603806988928137 1 -330 1 4.9927454 0.9696536 0.044458635792904468 1 +325 0 -3.7860875 0.032489616 0.04765095016422824 0 +326 1 3.6223297 0.9163445 0.12603797604758099 1 +330 1 4.9927444 0.96965355 0.044458724475427286 1 334 1 5.514736 0.97960865 0.029722574521978982 1 335 0 -5.935418 0.006225399 0.0090094253868573462 0 337 0 -5.4043503 0.009396286 0.013620064546924614 0 339 1 5.1161976 0.97236645 0.040427975332860647 1 -340 1 5.5803347 0.98060757 0.028252197193330195 1 +340 1 5.5803356 0.98060757 0.028252197193330195 1 341 0 -5.4043503 0.009396286 0.013620064546924614 0 342 0 -5.4523587 0.009053541 0.013120984925728642 0 345 0 -5.935418 0.006225399 0.0090094253868573462 0 -351 0 -4.9212914 0.013645148 0.019821329360883679 0 -356 1 -0.9127703 0.24064322 2.0550323292420534 0 +351 0 -4.921291 0.013645153 0.019821336171893227 0 +356 1 -0.9127703 0.24064319 2.0550325079118106 0 357 1 9.874601 0.9993099 0.0009959512596500891 1 359 1 4.653248 0.96079826 0.057694551223446147 1 -362 0 -3.5174994 0.03977293 0.0585524891762653 0 -363 0 -1.6840982 0.1478302 0.23078717543384142 0 -364 0 -4.9212914 0.013645148 0.019821329360883679 0 +362 0 -3.5174994 0.039772928 0.058552483579195597 0 +363 0 -1.6840982 0.14783019 0.23078715020667279 0 +364 0 -4.921291 0.013645153 0.019821336171893227 0 365 0 -5.186825 0.011117356 0.016128775943501653 0 367 1 9.040358 0.99867666 0.0019104437717653883 1 -369 0 -5.1140842 0.011759752 0.017066281948069518 0 +369 0 -5.1140847 0.011759749 0.017066276509657308 0 372 0 -3.9699688 0.02826512 0.041365340879266133 0 374 0 -4.4530277 0.0195541 0.028490070077080931 0 375 0 -5.935418 0.006225399 0.0090094253868573462 0 380 0 -5.935418 0.006225399 0.0090094253868573462 0 382 0 -3.5015903 0.040250305 0.059269898420679812 0 -385 0 -3.4273872 0.042550266 0.062731346866081117 0 +385 0 -3.4273868 0.042550277 0.062731363705997648 0 386 1 5.084236 0.97168756 0.041435602932387204 1 -390 0 -5.4467115 0.009093205 0.013178732083492451 0 +390 0 -5.446712 0.009093203 0.013178728015659433 0 393 0 -5.881093 0.006493533 0.0093987365367615445 0 394 0 -4.909887 0.013765578 0.019997487805960071 0 397 0 -4.48624 0.019062832 0.027767364472202984 0 @@ -547,112 +547,112 @@ Instance Label Score Probability Log-loss Assigned 401 0 -4.9692993 0.013149481 0.019096523087584735 0 402 0 -2.7191267 0.07173818 0.10739631960424616 0 403 0 -3.7908158 0.032373708 0.047478124251200779 0 -405 0 -5.669884 0.0076494967 0.011078316903702529 0 -407 0 -5.669884 0.0076494967 0.011078316903702529 0 -408 0 -3.5375085 0.039180238 0.057662270098084903 0 -410 0 -5.669884 0.0076494967 0.011078316903702529 0 +405 0 -5.669884 0.007649497 0.011078317580688359 0 +407 0 -5.669884 0.007649497 0.011078317580688359 0 +408 0 -3.537508 0.03918025 0.057662286878936264 0 +410 0 -5.669884 0.007649497 0.011078317580688359 0 411 0 ? ? ? 0 -412 1 7.6394253 0.996057 0.0056998288375171091 1 -417 0 -5.669884 0.0076494967 0.011078316903702529 0 -420 0 -2.6969714 0.07289933 0.10920208612242448 0 +412 1 7.639426 0.996057 0.0056998288375171091 1 +417 0 -5.669884 0.007649497 0.011078317580688359 0 +420 0 -2.696971 0.07289934 0.109202109310663 0 421 1 9.498289 0.9990743 0.0013361498208689703 1 424 0 -4.9692993 0.013149481 0.019096523087584735 0 425 1 11.849485 0.99985236 0.00021301623841437668 1 -426 0 -2.2324486 0.101552226 0.15449345077948776 0 +426 0 -2.2324486 0.10155222 0.15449343881561414 0 427 1 4.1596622 0.9433984 0.084060916290148127 1 -431 0 -2.9302087 0.06150328 0.091576391900387383 0 -432 0 -3.5334377 0.03930013 0.057842300709135597 0 +431 0 -2.9302087 0.061503276 0.09157638617372077 0 +432 0 -3.5334377 0.039300125 0.057842295114820461 0 433 0 -4.013695 0.02734187 0.039995280296142854 0 435 1 7.0216722 0.99362683 0.009223960722536808 1 437 0 -4.48624 0.019062832 0.027767364472202984 0 -438 0 -3.5384212 0.039153405 0.057621979836806482 0 +438 0 -3.5384207 0.039153416 0.057621996617189197 0 443 0 -5.54259 0.00844254 0.012231716526506063 0 -444 0 -2.6761093 0.074008495 0.11092913595470864 0 +444 0 -2.6761103 0.074008435 0.11092904309065733 0 445 0 -5.4523587 0.009053541 0.013120984925728642 0 446 0 -5.935418 0.006225399 0.0090094253868573462 0 -447 0 -4.0031805 0.027561162 0.040320580427227401 0 +447 0 -4.0031805 0.02756116 0.040320577663836212 0 448 0 -5.881093 0.006493533 0.0093987365367615445 0 458 0 -3.7524433 0.03332603 0.048898700380320723 0 -459 0 -3.5017061 0.04024681 0.059264645767511073 0 -460 0 -3.5568361 0.038615778 0.056814968504256366 0 -461 0 -3.3979883 0.04349581 0.064156808821778086 0 -462 0 -3.0737762 0.055339195 0.082131694177837 0 -463 0 -4.282031 0.022286426 0.032516212828799247 0 +459 0 -3.5017061 0.040246807 0.05926464016767781 0 +460 0 -3.5568361 0.038615774 0.056814962913923475 0 +461 0 -3.3979883 0.043495808 0.064156803202923621 0 +462 0 -3.0737767 0.05533917 0.082131654352744068 0 +463 0 -4.282031 0.022286424 0.032516210080316472 0 468 0 -4.48624 0.019062832 0.027767364472202984 0 469 0 -5.4191465 0.009289303 0.013464265265905247 0 -470 0 -4.649441 0.016819378 0.024471614226605987 0 -471 0 -3.0737762 0.055339195 0.082131694177837 0 +470 0 -4.649441 0.016819376 0.024471611493406355 0 +471 0 -3.0737767 0.05533917 0.082131654352744068 0 472 0 -3.4657965 0.04134449 0.060915614533985189 0 473 0 -4.48624 0.019062832 0.027767364472202984 0 475 0 -4.9692993 0.013149481 0.019096523087584735 0 -476 0 -4.2355022 0.023092356 0.033705916761290887 0 +476 0 -4.2355027 0.023092348 0.03370590575829005 0 477 0 -4.48624 0.019062832 0.027767364472202984 0 -478 0 -3.744658 0.03352252 0.049191978777848748 0 +478 0 -3.744658 0.033522516 0.049191973216976458 0 479 1 6.673233 0.9916496 0.012097621138980599 1 -481 0 -2.5658813 0.080130555 0.120498977234164 0 +481 0 -2.5658813 0.08013055 0.12049896554890183 0 485 0 -4.6490927 0.016823875 0.024478212185620931 0 -486 0 -4.649441 0.016819378 0.024471614226605987 0 -488 1 0.9591036 0.5776565 0.79171621850183493 1 +486 0 -4.649441 0.016819376 0.024471611493406355 0 +488 1 0.95910263 0.57765627 0.79171681395149796 1 490 0 -5.935418 0.006225399 0.0090094253868573462 0 -491 1 5.556223 0.98024607 0.028784146529073261 1 -494 0 -0.015696526 0.38975123 0.71253060259071122 0 +491 1 5.556222 0.98024607 0.028784146529073261 1 +494 0 -0.015696526 0.3897512 0.7125305321347537 0 496 0 -5.881093 0.006493533 0.0093987365367615445 0 -498 0 -4.220706 0.023354558 0.03409318909369452 0 -499 0 -4.220706 0.023354558 0.03409318909369452 0 -500 0 -2.9890537 0.05890274 0.087584266111585152 0 -503 0 -3.9551725 0.028584337 0.04183934627600109 0 +498 0 -4.220706 0.023354556 0.034093186342205803 0 +499 0 -4.220706 0.023354556 0.034093186342205803 0 +500 0 -2.9890532 0.058902755 0.087584288954953682 0 +503 0 -3.955172 0.028584346 0.041839360107510271 0 505 0 -4.4522543 0.019565688 0.028507120839304224 0 506 1 8.766861 0.99836195 0.0023651512585342507 1 -508 0 -4.0031805 0.027561162 0.040320580427227401 0 +508 0 -4.0031805 0.02756116 0.040320577663836212 0 509 0 -5.4523587 0.009053541 0.013120984925728642 0 511 0 -3.7376466 0.033700433 0.049457579363906538 0 -512 0 -4.0031805 0.027561162 0.040320580427227401 0 +512 0 -4.0031805 0.02756116 0.040320577663836212 0 515 1 7.3749876 0.9951566 0.0070045457965879953 1 516 0 -5.881093 0.006493533 0.0093987365367615445 0 -518 0 -4.49539 0.018929627 0.02757146870195281 0 +518 0 -4.4953895 0.018929632 0.027571476919188806 0 524 0 -4.7037654 0.016131794 0.023463022615813851 0 525 0 -4.746127 0.015614877 0.022705241025312198 0 526 0 -4.48624 0.019062832 0.027767364472202984 0 -530 1 5.325206 0.9764312 0.034409713804049308 1 -536 0 -3.4721131 0.041149348 0.060621972477918451 0 -537 0 -3.2213755 0.049611427 0.073410604725632031 0 -542 0 -3.4861355 0.0407193 0.059975064305642013 0 -543 0 -4.220706 0.023354558 0.03409318909369452 0 +530 1 5.325205 0.9764312 0.034409713804049308 1 +536 0 -3.4721127 0.041149363 0.060621994898335516 0 +537 0 -3.2213755 0.049611423 0.073410599070621044 0 +542 0 -3.4861355 0.040719297 0.059975058703050588 0 +543 0 -4.220706 0.023354556 0.034093186342205803 0 545 0 -3.7376466 0.033700433 0.049457579363906538 0 550 0 -4.7037654 0.016131794 0.023463022615813851 0 551 0 -5.4043503 0.009396286 0.013620064546924614 0 -552 0 -3.2360563 0.049073458 0.072594195629606423 0 -553 0 -1.4225526 0.17546201 0.27834212421314986 0 +552 0 -3.2360563 0.049073454 0.07259418997779464 0 +553 0 -1.4225531 0.17546193 0.27834199385026431 0 554 0 -4.9692993 0.013149481 0.019096523087584735 0 555 0 -1.7431297 0.14211455 0.22114307136971645 0 -556 0 -2.9508896 0.06057732 0.090153673385328204 0 +556 0 -2.9508896 0.060577318 0.090153667664306186 0 562 0 -5.4043503 0.009396286 0.013620064546924614 0 -564 0 -3.7509632 0.0333633 0.048954326671071656 0 -567 0 -3.4350505 0.04230704 0.062364900694131351 0 -568 1 3.5748348 0.913456 0.13059282121029733 1 +564 0 -3.7509632 0.033363298 0.04895432111111532 0 +567 0 -3.4350505 0.042307038 0.062364895082251487 0 +568 1 3.5748348 0.91345596 0.1305929153487409 1 570 1 6.466878 0.99020326 0.014203393857934222 1 -571 1 9.048693 0.99868524 0.0018980446658527542 1 +571 1 9.048691 0.99868524 0.0018980446658527542 1 572 0 -4.7037654 0.016131794 0.023463022615813851 0 -573 0 -5.669884 0.0076494967 0.011078316903702529 0 +573 0 -5.669884 0.007649497 0.011078317580688359 0 574 1 5.533702 0.9799025 0.029289877586105746 1 -575 0 -3.2213755 0.049611427 0.073410604725632031 0 +575 0 -3.2213755 0.049611423 0.073410599070621044 0 576 0 -3.7376466 0.033700433 0.049457579363906538 0 579 0 -5.4043503 0.009396286 0.013620064546924614 0 580 0 -3.4869094 0.04069569 0.059939555518058511 0 583 0 -4.9692993 0.013149481 0.019096523087584735 0 585 0 -5.935418 0.006225399 0.0090094253868573462 0 -587 0 -3.5334377 0.03930013 0.057842300709135597 0 +587 0 -3.5334377 0.039300125 0.057842295114820461 0 588 1 4.6442137 0.9605316 0.058095029247836778 1 -589 0 -4.0031805 0.027561162 0.040320580427227401 0 -591 1 4.243066 0.9467787 0.078900822913083493 1 -592 1 4.8517914 0.96624035 0.049545999536437443 1 +589 0 -4.0031805 0.02756116 0.040320577663836212 0 +591 1 4.243067 0.9467788 0.078900732087929318 1 +592 1 4.8517895 0.9662403 0.049546088532232753 1 595 0 -3.7376466 0.033700433 0.049457579363906538 0 596 0 -3.9699688 0.02826512 0.041365340879266133 0 -597 0 -2.9706383 0.059705295 0.08881510055974505 0 +597 0 -2.9706383 0.05970529 0.088815094844028691 0 598 0 -4.7037654 0.016131794 0.023463022615813851 0 -599 0 -2.9381208 0.061147474 0.091029535901308245 0 +599 0 -2.9381208 0.06114747 0.091029530176811924 0 601 0 -5.6155596 0.007978473 0.011556667054376319 0 -603 1 3.1762495 0.8854623 0.17549723761975466 1 +603 1 3.1762505 0.8854623 0.17549723761975466 1 605 1 8.159748 0.99737054 0.0037985035173251067 1 608 1 8.079367 0.9972006 0.0040443320699722306 1 610 1 6.972576 0.9933793 0.0095834180703139316 1 @@ -661,38 +661,38 @@ Instance Label Score Probability Log-loss Assigned 616 0 -4.7037654 0.016131794 0.023463022615813851 0 620 0 -4.7037654 0.016131794 0.023463022615813851 0 623 0 -5.935418 0.006225399 0.0090094253868573462 0 -625 0 -3.343666 0.04529595 0.066874515461359224 0 -626 1 3.8647957 0.92976606 0.10506033360236412 1 +625 0 -3.343666 0.045295946 0.066874509831910151 0 +626 1 3.8647957 0.929766 0.10506042608942295 1 628 0 -5.4523587 0.009053541 0.013120984925728642 0 -630 0 -2.7601237 0.06963448 0.10413046788911323 0 +630 0 -2.7601237 0.069634475 0.10413045633568027 0 631 0 -3.7376466 0.033700433 0.049457579363906538 0 632 0 -5.935418 0.006225399 0.0090094253868573462 0 -635 0 -4.217087 0.023419132 0.034188580858367221 0 -636 1 8.162584 0.9973763 0.0037901403924413767 1 -637 0 -2.370799 0.09210701 0.13940583705428458 0 -640 0 -4.0398955 0.026802769 0.039195879346332799 0 +635 0 -4.2170873 0.023419123 0.034188567100013993 0 +636 1 8.162586 0.9973763 0.0037901403924413767 1 +637 0 -2.370799 0.092107005 0.13940582521487643 0 +640 0 -4.039895 0.026802778 0.039195893152521553 0 643 0 -5.935418 0.006225399 0.0090094253868573462 0 646 0 -5.426158 0.00923903 0.013391058760089801 0 647 0 -5.4890733 0.008799812 0.012751633519735472 0 -648 1 8.579458 0.9981042 0.0027376359023654488 1 -650 0 -3.6219687 0.036770523 0.054048551776378198 0 +648 1 8.579456 0.9981042 0.0027376359023654488 1 +650 0 -3.6219683 0.036770534 0.054048568515248828 0 651 0 -4.965017 0.013192966 0.019160095111675445 0 655 0 -4.7037654 0.016131794 0.023463022615813851 0 658 1 7.546403 0.99576104 0.006128529583862446 1 659 0 -5.935418 0.006225399 0.0090094253868573462 0 662 0 -5.1684093 0.011276633 0.016361165642268163 0 663 0 -5.1684093 0.011276633 0.016361165642268163 0 -664 0 -4.2574205 0.022709223 0.03314021761930909 0 -666 0 -3.070212 0.05548494 0.082354293797361033 0 +664 0 -4.2574205 0.02270922 0.033140214869637266 0 +666 0 -3.0702114 0.055484954 0.08235431655806906 0 667 0 -4.4382315 0.01977694 0.028818008476875413 0 -669 1 6.981786 0.99342644 0.0095149472206866505 1 -671 0 -3.9565368 0.02855476 0.04179542083805643 0 -672 0 -4.9212914 0.013645148 0.019821329360883679 0 -673 0 -3.289723 0.04715361 0.06968444246223128 0 -674 0 -5.669884 0.0076494967 0.011078316903702529 0 -675 0 -3.7340279 0.033792615 0.049595215049171026 0 +669 1 6.9817867 0.99342644 0.0095149472206866505 1 +671 0 -3.9565368 0.028554758 0.041795418071838834 0 +672 0 -4.921291 0.013645153 0.019821336171893227 0 +673 0 -3.289723 0.047153607 0.069684436821807083 0 +674 0 -5.669884 0.007649497 0.011078317580688359 0 +675 0 -3.7340279 0.03379261 0.049595209486744243 0 676 0 -5.4191465 0.009289303 0.013464265265905247 0 -677 0 -4.0031805 0.027561162 0.040320580427227401 0 +677 0 -4.0031805 0.02756116 0.040320577663836212 0 684 0 -5.935418 0.006225399 0.0090094253868573462 0 686 0 -5.935418 0.006225399 0.0090094253868573462 0 687 0 -4.295347 0.022060877 0.032183434683621268 0 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration-out.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration-out.txt index 9206366946..45cc4803b0 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration-out.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration-out.txt @@ -13,15 +13,15 @@ Confusion table ||====================== PREDICTED || positive | negative | Recall TRUTH ||====================== - positive || 131 | 3 | 0.9776 - negative || 8 | 212 | 0.9636 + positive || 133 | 1 | 0.9925 + negative || 9 | 211 | 0.9591 ||====================== -Precision || 0.9424 | 0.9860 | -OVERALL 0/1 ACCURACY: 0.968927 +Precision || 0.9366 | 0.9953 | +OVERALL 0/1 ACCURACY: 0.971751 LOG LOSS/instance: NaN Test-set entropy (prior Log-Loss/instance): 0.956998 LOG-LOSS REDUCTION (RIG): 0.000000 -AUC: 0.994437 +AUC: 0.994403 Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. TEST POSITIVE RATIO: 0.3191 (105.0/(105.0+224.0)) Confusion table @@ -40,16 +40,16 @@ AUC: 0.997619 OVERALL RESULTS --------------------------------------- -AUC: 0.996028 (0.0016) -Accuracy: 0.972305 (0.0034) -Positive precision: 0.956660 (0.0142) -Positive recall: 0.964996 (0.0126) -Negative precision: 0.981961 (0.0041) -Negative recall: 0.975122 (0.0115) +AUC: 0.996011 (0.0016) +Accuracy: 0.973718 (0.0020) +Positive precision: 0.953747 (0.0171) +Positive recall: 0.972459 (0.0201) +Negative precision: 0.986580 (0.0087) +Negative recall: 0.972849 (0.0138) Log-loss: NaN (NaN) Log-loss reduction: 0.000000 (0.0000) -F1 Score: 0.960623 (0.0009) -AUPRC: 0.992280 (0.0025) +F1 Score: 0.962653 (0.0011) +AUPRC: 0.992269 (0.0025) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration-rp.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration-rp.txt index d0fd8a6afd..9e9d966393 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration-rp.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration-rp.txt @@ -1,4 +1,4 @@ AveragedPerceptron AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings -0.996028 0.972305 0.95666 0.964996 0.981961 0.975122 NaN 0 0.960623 0.99228 AveragedPerceptron %Data% %Output% 99 0 0 maml.exe CV tr=AveragedPerceptron threads=- cali={} dout=%Output% data=%Data% seed=1 +0.996011 0.973718 0.953747 0.972459 0.98658 0.972849 NaN 0 0.962653 0.992269 AveragedPerceptron %Data% %Output% 99 0 0 maml.exe CV tr=AveragedPerceptron threads=- cali={} dout=%Output% data=%Data% seed=1 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration.txt index d5dee0b8f6..8ee0e5ef80 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-CV-breast-cancer.nocalibration.txt @@ -1,379 +1,379 @@ Instance Label Score Assigned -5 1 11.285979 1 -6 0 -0.93471193 0 -8 0 -3.7769966 0 -9 0 -3.7947202 0 -10 0 -4.7855167 0 -11 0 -4.6251884 0 -18 1 6.8803635 1 -20 1 5.573552 1 -21 1 6.7444105 1 -25 1 1.2789736 1 -28 0 -4.6251884 0 -31 0 -4.3083706 0 -32 1 6.9428844 1 -35 0 -4.6251884 0 -37 0 -1.8143315 0 +5 1 11.925824 1 +6 0 -0.4527979 0 +8 0 -3.7962546 0 +9 0 -3.8130417 0 +10 0 -4.7285223 0 +11 0 -4.606969 0 +18 1 7.3475924 1 +20 1 6.1389017 1 +21 1 7.1486177 1 +25 1 1.6632223 1 +28 0 -4.606969 0 +31 0 -4.2645645 0 +32 1 7.198575 1 +35 0 -4.606969 0 +37 0 -1.714282 0 40 0 ? 0 -41 1 2.4075565 1 -44 1 8.039285 1 -45 0 -4.625085 0 -46 1 5.138131 1 -48 0 -3.4336777 0 -50 1 2.7120514 1 -51 1 -0.062075615 0 -52 1 4.4027233 1 -54 1 6.3079214 1 -56 1 6.356518 1 -60 1 1.9474735 1 -63 1 0.78555584 1 -64 0 -4.916337 0 -66 0 -3.7260728 0 -68 1 9.2772875 1 -69 0 -4.4157114 0 -70 0 -3.0868545 0 -71 1 7.5159607 1 -72 0 -1.8410158 0 -73 1 7.1320066 1 -74 1 2.4329157 1 -76 0 -3.9190063 0 -77 0 -3.1092033 0 -79 0 -4.4391913 0 -82 0 -3.1867537 0 -88 0 -3.7260728 0 -90 0 -4.5995197 0 -91 0 -4.5046597 0 -92 0 -3.7260728 0 -93 0 -4.916337 0 -95 0 -4.5995197 0 -96 0 -4.7958083 0 -97 0 -3.4349236 0 -98 1 9.07517 1 -99 1 8.952344 1 -100 1 4.9092436 1 -102 0 -3.393629 0 -104 1 10.959611 1 -105 1 2.0113592 1 -106 1 8.251353 1 -108 0 -4.5487204 0 -109 1 5.864868 1 -111 1 3.7846975 1 -112 1 6.380026 1 -113 1 9.461209 1 -115 0 -3.6043515 0 -117 1 7.9902315 1 -120 0 -4.120878 0 -121 0 -3.0707016 0 -122 1 10.129083 1 -123 1 4.173232 1 -125 0 -4.916337 0 -128 1 4.5026884 1 -129 0 -3.7451797 0 -131 0 -4.3083706 0 -132 1 8.723828 1 -133 0 -4.150231 0 -137 0 -4.650857 0 -138 0 -3.7104468 0 -141 0 -4.942006 0 -144 0 -4.6251884 0 +41 1 2.5451756 1 +44 1 8.165841 1 +45 0 -4.602255 0 +46 1 5.6216097 1 +48 0 -3.379683 0 +50 1 2.8003244 1 +51 1 0.14775372 1 +52 1 4.696246 1 +54 1 6.743867 1 +56 1 6.5947084 1 +60 1 2.2064123 1 +63 1 0.8789625 1 +64 0 -4.8905344 0 +66 0 -3.697434 0 +68 1 9.899808 1 +69 0 -4.3595524 0 +70 0 -3.0557137 0 +71 1 7.555622 1 +72 0 -1.676908 0 +73 1 7.7111263 1 +74 1 2.4994192 1 +76 0 -3.839695 0 +77 0 -3.1438046 0 +79 0 -4.4265766 0 +82 0 -3.1870723 0 +88 0 -3.697434 0 +90 0 -4.54813 0 +91 0 -4.5069323 0 +92 0 -3.697434 0 +93 0 -4.8905344 0 +95 0 -4.54813 0 +96 0 -4.790498 0 +97 0 -3.413869 0 +98 1 9.294541 1 +99 1 9.62131 1 +100 1 5.1074305 1 +102 0 -3.3471546 0 +104 1 11.120679 1 +105 1 2.1430416 1 +106 1 8.747506 1 +108 0 -4.513379 0 +109 1 6.3912535 1 +111 1 4.1208715 1 +112 1 7.006652 1 +113 1 9.855809 1 +115 0 -3.4127908 0 +117 1 8.638749 1 +120 0 -4.038975 0 +121 0 -3.0156002 0 +122 1 10.533511 1 +123 1 4.6804914 1 +125 0 -4.8905344 0 +128 1 4.9510326 1 +129 0 -3.7898016 0 +131 0 -4.2645645 0 +132 1 9.206267 1 +133 0 -4.134826 0 +137 0 -4.6658077 0 +138 0 -3.6895585 0 +141 0 -4.9493732 0 +144 0 -4.606969 0 145 0 ? 0 -147 0 -4.3231397 0 -150 0 -4.7855167 0 -151 1 4.569235 1 -152 1 8.551608 1 -154 0 -5.233155 0 -156 0 -4.3357234 0 -161 0 -3.5422645 0 +147 0 -4.255105 0 +150 0 -4.7285223 0 +151 1 5.0748987 1 +152 1 8.881612 1 +154 0 -5.232939 0 +156 0 -4.3134584 0 +161 0 -3.5088563 0 164 0 ? 0 -167 1 7.4310427 1 -169 0 -5.2729545 0 -171 0 -4.5995197 0 -173 1 13.79783 1 -174 1 5.2489986 1 -176 0 -4.3083706 0 -177 1 6.069394 1 -179 1 2.427494 1 -180 0 -4.7855167 0 -181 0 -5.233155 0 -183 1 8.429484 1 -187 1 11.391686 1 -188 1 7.5789557 1 -189 0 -3.686462 0 -191 1 10.154208 1 -192 0 -3.7517414 0 -196 0 5.425536 1 -198 0 -5.233155 0 -199 0 -4.334039 0 -201 1 9.260581 1 -202 0 -4.5995197 0 -204 0 -4.5995197 0 -205 1 11.455533 1 -206 1 6.5003185 1 -207 0 -4.7855167 0 -209 0 -3.6209207 0 -210 1 12.897248 1 -211 1 8.615066 1 -212 0 -4.5995197 0 -216 0 -4.916337 0 -218 1 7.456811 1 -219 0 -2.4788876 0 -223 1 4.949767 1 -226 1 8.538112 1 -228 0 -4.7855167 0 -233 1 5.219512 1 -237 1 6.374799 1 -239 1 4.7197256 1 -240 0 -2.0196419 0 -241 0 -3.9108233 0 -242 0 -4.3083706 0 -244 0 -4.5995197 0 -246 1 10.46416 1 -247 1 2.8055887 1 -248 0 -2.8070307 0 +167 1 7.490013 1 +169 0 -5.254455 0 +171 0 -4.54813 0 +173 1 14.494651 1 +174 1 5.7514915 1 +176 0 -4.2645645 0 +177 1 6.138485 1 +179 1 2.8749352 1 +180 0 -4.7285223 0 +181 0 -5.232939 0 +183 1 9.06396 1 +187 1 12.00492 1 +188 1 8.132482 1 +189 0 -3.6207743 0 +191 1 10.797475 1 +192 0 -3.7562728 0 +196 0 5.8725023 1 +198 0 -5.232939 0 +199 0 -4.3234034 0 +201 1 9.770606 1 +202 0 -4.54813 0 +204 0 -4.54813 0 +205 1 12.087651 1 +206 1 6.641532 1 +207 0 -4.7285223 0 +209 0 -3.5942607 0 +210 1 13.547517 1 +211 1 9.089206 1 +212 0 -4.54813 0 +216 0 -4.8905344 0 +218 1 7.7499723 1 +219 0 -2.4297438 0 +223 1 5.4305964 1 +226 1 9.166205 1 +228 0 -4.7285223 0 +233 1 5.6998796 1 +237 1 6.476473 1 +239 1 5.1697493 1 +240 0 -1.9057703 0 +241 0 -3.8436403 0 +242 0 -4.2645645 0 +244 0 -4.54813 0 +246 1 11.068807 1 +247 1 3.1855059 1 +248 0 -2.7545462 0 249 0 ? 0 -250 0 -4.652541 0 -252 0 3.7707853 1 -254 1 7.661195 1 -257 0 -4.334039 0 -258 0 -4.0172215 0 -259 0 4.008581 1 -260 1 7.7233286 1 -262 1 9.050584 1 -267 1 3.073165 1 -268 1 8.364619 1 -269 0 -4.5995197 0 -271 0 -3.4349236 0 -272 1 3.073165 1 +250 0 -4.655863 0 +252 0 4.2659445 1 +254 1 7.919627 1 +257 0 -4.3234034 0 +258 0 -3.9809995 0 +259 0 4.3407555 1 +260 1 8.330071 1 +262 1 9.629434 1 +267 1 3.677929 1 +268 1 8.609313 1 +269 0 -4.54813 0 +271 0 -3.413869 0 +272 1 3.677929 1 275 0 ? 0 -276 0 -4.334039 0 -277 0 -4.916337 0 -278 0 -4.5995197 0 -279 1 6.3361444 1 -280 0 -4.0172215 0 -283 1 5.09361 1 -284 1 5.780157 1 -285 1 12.663693 1 -288 1 1.8098211 1 -290 0 -5.233155 0 -291 0 -4.5995197 0 -293 1 4.209258 1 -296 0 1.7716074 1 +276 0 -4.3234034 0 +277 0 -4.8905344 0 +278 0 -4.54813 0 +279 1 6.636298 1 +280 0 -3.9809995 0 +283 1 5.5930414 1 +284 1 6.3683043 1 +285 1 13.342931 1 +288 1 2.133194 1 +290 0 -5.232939 0 +291 0 -4.54813 0 +293 1 4.659646 1 +296 0 2.0631151 1 297 0 ? 0 -299 1 5.4912786 1 -300 1 6.2749596 1 -301 0 -4.5995197 0 -303 0 -4.5995197 0 -304 1 4.320197 1 -308 1 6.5411158 1 -309 0 -1.7547832 0 -311 0 -5.233155 0 -312 1 3.2156925 1 -314 0 -5.102334 0 -316 1 3.9409618 1 -317 1 8.260409 1 -319 0 2.46204 1 +299 1 6.110527 1 +300 1 6.6979456 1 +301 0 -4.54813 0 +303 0 -4.54813 0 +304 1 4.865142 1 +308 1 6.804652 1 +309 0 -1.4760432 0 +311 0 -5.232939 0 +312 1 3.2489738 1 +314 0 -5.070926 0 +316 1 4.409379 1 +317 1 8.798545 1 +319 0 2.7989092 1 321 0 ? 0 -323 1 4.269208 1 -327 0 -4.916337 0 -328 1 4.1030073 1 -329 1 6.3562107 1 -331 0 -3.1436715 0 -332 0 -2.8411217 0 -333 1 4.4240284 1 -336 1 4.790291 1 -338 0 -5.102334 0 -343 0 -5.233155 0 -344 1 8.780366 1 -346 0 -2.788134 0 -347 0 -5.0515347 0 -348 1 -0.033994675 0 -349 1 2.9944906 1 -350 0 -3.776164 0 -352 0 1.3297043 1 -353 1 8.744256 1 -354 0 -4.916337 0 -355 0 -3.6730852 0 -358 1 6.1854076 1 -360 1 14.4099455 1 -361 1 6.113164 1 -366 1 12.847375 1 -368 0 -4.568268 0 -370 0 -3.0281096 0 -371 0 -4.568268 0 -373 0 -3.605544 0 -376 0 -4.916337 0 -377 0 -5.102334 0 -378 0 -3.5256414 0 -379 0 -1.5692582 0 -381 1 8.122036 1 -383 0 -4.942006 0 -384 0 -4.942006 0 -387 0 -1.9415016 0 -388 0 -4.44138 0 -389 0 -2.9747353 0 -391 1 8.779809 1 -392 0 -4.334039 0 -395 0 -4.334039 0 -396 0 -4.0172215 0 -398 0 -3.902061 0 -399 0 -4.3200974 0 -404 0 -4.508781 0 -406 0 -3.4622765 0 -409 0 -4.0015955 0 -413 0 -3.1024804 0 -414 1 5.959919 1 -415 0 -0.721817 0 -416 1 8.443301 1 -418 0 -1.8258505 0 -419 0 -3.8876748 0 -422 0 -2.1972284 0 -423 0 -3.0868545 0 -428 0 -4.916337 0 -429 0 -4.6251884 0 -430 0 -4.2361884 0 -434 0 5.330062 1 -436 1 4.9601755 1 -439 0 -4.068559 0 -440 1 7.0005217 1 -441 0 -1.8277493 0 -442 0 -4.126358 0 -449 1 9.384189 1 -450 0 -3.936492 0 -451 0 -4.068559 0 -452 0 -4.358462 0 -453 1 7.3491344 1 -454 0 -4.259669 0 -455 1 0.2950573 1 -456 1 8.340758 1 -457 1 7.996641 1 -464 0 -4.3597083 0 -465 1 8.680116 1 -466 1 8.110646 1 -467 1 6.858451 1 -474 0 -4.068559 0 -480 0 -4.254556 0 -482 1 13.881022 1 -483 1 9.617421 1 -484 0 -3.736116 0 -487 1 11.720016 1 -489 1 -0.5905738 0 -492 0 -4.228887 0 -493 1 9.492114 1 -495 0 -4.520036 0 -497 0 -4.111538 0 -501 0 -4.0428905 0 -502 0 -3.8966928 0 -504 0 -5.233155 0 -507 0 -4.0115523 0 -510 0 -5.233155 0 -513 0 -4.520036 0 -514 1 8.787938 1 -517 0 -5.102334 0 -519 1 6.320156 1 -520 0 -5.0471582 0 -521 0 -4.1737113 0 -522 1 3.983387 1 -523 1 6.156104 1 -527 0 -3.7260728 0 -528 0 -2.9663248 0 -529 0 -4.228887 0 -531 0 -3.4622765 0 -532 0 -4.7855167 0 -533 0 -4.334039 0 -534 0 -4.6251884 0 -535 0 -3.7884345 0 -538 0 -4.0428905 0 -539 0 -3.4605927 0 -540 0 -3.3445406 0 -541 0 -4.650857 0 -544 0 -3.8141036 0 -546 1 10.355874 1 -547 0 -5.128003 0 -548 0 -4.836854 0 -549 1 5.2726183 1 -557 0 -3.776164 0 -558 0 -4.6251884 0 -559 0 -3.7517414 0 -560 0 -3.4349236 0 -561 0 -3.4349236 0 -563 0 -4.334039 0 -565 1 10.456666 1 -566 0 -3.6847782 0 -569 1 8.855367 1 -577 0 -4.916337 0 -578 0 -4.916337 0 -581 1 8.00238 1 -582 1 7.645852 1 -584 0 -3.1515741 0 -586 1 12.260621 1 -590 1 4.0090714 1 -593 0 -3.736116 0 -594 1 5.269803 1 -600 0 -4.334039 0 -602 0 -4.0428905 0 -604 1 4.515382 1 -606 0 -4.2135105 0 -607 0 -5.233155 0 -609 0 -4.068559 0 -612 1 14.881892 1 -613 0 -4.128848 0 -614 0 -4.8111854 0 +323 1 4.8855343 1 +327 0 -4.8905344 0 +328 1 4.1213064 1 +329 1 6.918497 1 +331 0 -3.125589 0 +332 0 -2.8317442 0 +333 1 4.9879713 1 +336 1 5.3119774 1 +338 0 -5.070926 0 +343 0 -5.232939 0 +344 1 9.373975 1 +346 0 -2.8051786 0 +347 0 -5.0361757 0 +348 1 0.09843922 1 +349 1 3.543579 1 +350 0 -3.7809267 0 +352 0 1.5155859 1 +353 1 9.391765 1 +354 0 -4.8905344 0 +355 0 -3.6708684 0 +358 1 6.2604885 1 +360 1 15.035466 1 +361 1 6.5571547 1 +366 1 13.467439 1 +368 0 -4.532379 0 +370 0 -3.0555735 0 +371 0 -4.532379 0 +373 0 -3.5973978 0 +376 0 -4.8905344 0 +377 0 -5.070926 0 +378 0 -3.4778075 0 +379 0 -1.5486526 0 +381 1 8.643491 1 +383 0 -4.9493732 0 +384 0 -4.9493732 0 +387 0 -1.9204488 0 +388 0 -4.418391 0 +389 0 -2.9322662 0 +391 1 9.323798 1 +392 0 -4.3234034 0 +395 0 -4.3234034 0 +396 0 -3.9809995 0 +398 0 -3.9080298 0 +399 0 -4.305583 0 +404 0 -4.4998136 0 +406 0 -3.4627619 0 +409 0 -3.973124 0 +413 0 -3.063589 0 +414 1 6.5001116 1 +415 0 -0.5379734 0 +416 1 8.981729 1 +418 0 -1.7460232 0 +419 0 -3.9546041 0 +422 0 -2.1661267 0 +423 0 -3.0557137 0 +428 0 -4.8905344 0 +429 0 -4.606969 0 +430 0 -4.248364 0 +434 0 5.678664 1 +436 1 4.8968496 1 +439 0 -4.098677 0 +440 1 7.6677313 1 +441 0 -1.6798639 0 +442 0 -4.0051136 0 +449 1 10.087396 1 +450 0 -3.9024792 0 +451 0 -4.098677 0 +452 0 -4.348057 0 +453 1 8.068194 1 +454 0 -4.315388 0 +455 1 0.4596901 1 +456 1 8.942605 1 +457 1 8.346686 1 +464 0 -4.382242 0 +465 1 8.953591 1 +466 1 8.690221 1 +467 1 7.4474783 1 +474 0 -4.098677 0 +480 0 -4.279069 0 +482 1 14.295785 1 +483 1 10.273198 1 +484 0 -3.7483978 0 +487 1 12.309564 1 +489 1 -0.43852377 0 +492 0 -4.22023 0 +493 1 10.096398 1 +495 0 -4.5037956 0 +497 0 -4.155446 0 +501 0 -4.039838 0 +502 0 -3.8809624 0 +504 0 -5.232939 0 +507 0 -3.8678274 0 +510 0 -5.232939 0 +513 0 -4.5037956 0 +514 1 9.508458 1 +517 0 -5.070926 0 +519 1 6.943595 1 +520 0 -5.0525465 0 +521 0 -4.2018504 0 +522 1 4.3094034 1 +523 1 6.7598047 1 +527 0 -3.697434 0 +528 0 -2.955677 0 +529 0 -4.22023 0 +531 0 -3.4627619 0 +532 0 -4.7285223 0 +533 0 -4.3234034 0 +534 0 -4.606969 0 +535 0 -3.6886954 0 +538 0 -4.039838 0 +539 0 -3.4727077 0 +540 0 -3.3012362 0 +541 0 -4.6658077 0 +544 0 -3.7475338 0 +546 1 11.016191 1 +547 0 -5.1297655 0 +548 0 -4.8462 0 +549 1 5.5535192 1 +557 0 -3.7809267 0 +558 0 -4.606969 0 +559 0 -3.7562728 0 +560 0 -3.413869 0 +561 0 -3.413869 0 +563 0 -4.3234034 0 +565 1 11.259841 1 +566 0 -3.6307197 0 +569 1 9.242663 1 +577 0 -4.8905344 0 +578 0 -4.8905344 0 +581 1 8.654239 1 +582 1 8.128011 1 +584 0 -3.123908 0 +586 1 12.971318 1 +590 1 4.1888075 1 +593 0 -3.7483978 0 +594 1 5.7850237 1 +600 0 -4.3234034 0 +602 0 -4.039838 0 +604 1 5.138941 1 +606 0 -4.2233667 0 +607 0 -5.232939 0 +609 0 -4.098677 0 +612 1 15.565835 1 +613 0 -4.153376 0 +614 0 -4.787361 0 617 0 ? 0 -618 0 -4.0428905 0 -619 0 -3.7517414 0 -621 0 -0.14650154 0 -622 0 -2.5446053 0 -624 0 -3.7915406 0 -627 0 -3.3132024 0 -629 0 -4.3597083 0 -633 1 4.0237722 1 -634 0 -4.650857 0 -638 0 -4.3597083 0 -639 0 -3.776164 0 -641 0 -4.334039 0 -642 0 -4.334039 0 -644 0 -4.942006 0 -645 0 -4.334039 0 -649 0 -4.334039 0 -652 0 -3.567933 0 -653 0 -4.0428905 0 -654 0 -4.0172215 0 -656 0 -3.7517414 0 -657 0 0.674386 1 -660 0 -4.916337 0 -661 0 -3.7260728 0 -665 0 -5.233155 0 -668 1 3.2994661 1 -670 1 6.4614477 1 -678 0 -5.233155 0 -679 0 -4.942006 0 -680 1 14.404435 1 -681 1 9.278363 1 -682 0 -3.2511153 0 -683 0 -5.233155 0 -685 0 -5.233155 0 -688 0 -4.3597083 0 -689 0 -3.1943884 0 -691 1 5.2444315 1 -692 0 -4.650857 0 -693 0 -4.042787 0 -694 0 -4.057659 0 -696 1 7.3569994 1 -697 1 4.656295 1 -698 1 5.6929607 1 -0 0 -3.4721131 0 +618 0 -4.039838 0 +619 0 -3.7562728 0 +621 0 0.0042414665 1 +622 0 -2.5211115 0 +624 0 -3.7777896 0 +627 0 -3.1292257 0 +629 0 -4.382242 0 +633 1 4.158328 1 +634 0 -4.6658077 0 +638 0 -4.382242 0 +639 0 -3.7809267 0 +641 0 -4.3234034 0 +642 0 -4.3234034 0 +644 0 -4.9493732 0 +645 0 -4.3234034 0 +649 0 -4.3234034 0 +652 0 -3.5676956 0 +653 0 -4.039838 0 +654 0 -3.9809995 0 +656 0 -3.7562728 0 +657 0 0.6928787 1 +660 0 -4.8905344 0 +661 0 -3.697434 0 +665 0 -5.232939 0 +668 1 3.362691 1 +670 1 6.9357576 1 +678 0 -5.232939 0 +679 0 -4.9493732 0 +680 1 15.087735 1 +681 1 9.870067 1 +682 0 -3.2252913 0 +683 0 -5.232939 0 +685 0 -5.232939 0 +688 0 -4.382242 0 +689 0 -3.214981 0 +691 1 5.504386 1 +692 0 -4.6658077 0 +693 0 -4.035124 0 +694 0 -4.030379 0 +696 1 7.6070995 1 +697 1 5.0033865 1 +698 1 6.1210947 1 +0 0 -3.4721127 0 1 0 2.4163914 1 -2 0 -4.045404 0 -3 0 2.9251795 1 +2 0 -4.0454035 0 +3 0 2.9251804 1 4 0 -3.5088277 0 -7 0 -4.670553 0 +7 0 -4.6705537 0 12 1 -0.34343147 0 -13 0 -4.6186943 0 +13 0 -4.618695 0 14 1 7.360214 1 -15 1 0.6494303 1 +15 1 0.6494312 1 16 0 -4.220706 0 -17 0 -3.9551725 0 -19 0 -2.9890537 0 +17 0 -3.955172 0 +19 0 -2.9890532 0 22 0 -4.7037654 0 23 1 ? 0 24 0 -5.4043503 0 @@ -381,47 +381,47 @@ Instance Label Score Assigned 27 0 -3.7376466 0 29 0 -5.4339433 0 30 0 -4.649441 0 -33 0 -4.698118 0 +33 0 -4.6981187 0 34 0 -4.4530277 0 36 1 7.832773 1 38 1 4.92861 1 39 1 1.079258 1 42 1 6.8985863 1 -43 1 -0.49528694 0 +43 1 -0.49528742 0 47 0 -5.669884 0 -49 1 5.3024044 1 +49 1 5.3024054 1 53 1 5.116103 1 55 1 4.4195347 1 -57 1 0.5701313 1 -58 1 1.1371031 1 +57 1 0.57013035 1 +58 1 1.137104 1 59 1 1.6442327 1 61 0 -5.2770567 0 -62 1 5.7670774 1 -65 1 2.7867746 1 +62 1 5.7670784 1 +65 1 2.7867756 1 67 1 2.9058514 1 -75 0 -4.291042 0 +75 0 -4.2910423 0 78 0 -3.652576 0 80 0 -3.2955709 0 81 0 -3.9975338 0 -83 0 -2.9832687 0 -84 1 6.5824003 1 +83 0 -2.9832683 0 +84 1 6.5823994 1 85 1 4.7604074 1 86 1 1.466999 1 -87 1 5.223543 1 +87 1 5.223544 1 89 0 -5.017977 0 -94 0 -4.9212914 0 -101 1 -0.85990286 0 +94 0 -4.921291 0 +101 1 -0.85990334 0 103 1 0.36755466 1 107 1 4.617798 1 110 0 -3.1312823 0 114 0 -2.8847933 0 -116 0 -0.40737772 0 +116 0 -0.40737724 0 118 0 -5.2584124 0 -119 0 -3.786881 0 +119 0 -3.7868814 0 124 1 5.0384197 1 126 1 6.4765244 1 127 0 -4.4382315 0 -130 0 -3.2361722 0 +130 0 -3.2361717 0 134 0 -4.75809 0 135 0 -2.670867 0 136 0 -4.220706 0 @@ -429,12 +429,12 @@ Instance Label Score Assigned 140 0 -4.9692993 0 142 1 3.7043686 1 143 0 -4.364107 0 -146 1 0.36148834 1 -148 0 -2.357305 0 -149 1 8.756336 1 +146 1 0.3614874 1 +148 0 -2.3573046 0 +149 1 8.756334 1 153 0 -3.7005844 0 155 1 3.1619148 1 -157 0 -4.9212914 0 +157 0 -4.921291 0 158 0 ? 0 159 1 10.452139 1 160 1 7.997595 1 @@ -445,62 +445,62 @@ Instance Label Score Assigned 168 0 -4.4382315 0 170 0 -4.9692993 0 172 0 -5.669884 0 -175 1 6.0381765 1 -178 0 -3.9551725 0 -182 0 -2.9890537 0 +175 1 6.0381775 1 +178 0 -3.955172 0 +182 0 -2.9890532 0 184 1 5.409273 1 185 0 -4.8669662 0 186 1 3.9876003 1 190 1 10.521242 1 193 0 -5.4043503 0 194 0 -4.4382315 0 -195 0 -3.9551725 0 +195 0 -3.955172 0 197 0 -2.8025956 0 200 1 8.673521 1 -203 0 -3.4721131 0 -208 0 -5.350025 0 +203 0 -3.4721127 0 +208 0 -5.3500257 0 213 1 12.300528 1 214 1 11.895983 1 -215 1 6.600219 1 +215 1 6.6002197 1 217 0 -5.4043503 0 220 0 -5.181178 0 221 1 7.9662914 1 222 1 -2.1487255 0 -224 1 8.4735565 1 +224 1 8.473555 1 225 0 -5.669884 0 227 1 6.748211 1 229 1 10.504805 1 -230 1 4.829337 1 +230 1 4.829338 1 231 1 6.912092 1 232 0 1.0722923 1 234 0 -2.7037287 0 235 0 ? 0 236 1 9.440506 1 238 1 10.690645 1 -243 0 -3.301972 0 -245 0 -2.856113 0 +243 0 -3.3019714 0 +245 0 -2.8561125 0 251 1 7.355525 1 253 1 6.8985863 1 -255 1 3.745204 1 +255 1 3.745205 1 256 0 -4.9692993 0 261 1 9.009869 1 263 1 7.1455708 1 264 1 4.0145664 1 -265 0 -2.5156913 0 +265 0 -2.5156918 0 266 1 7.325534 1 270 1 5.5723915 1 -273 1 0.03731823 1 +273 1 0.037317276 1 274 0 -4.2340226 0 -281 0 -4.698118 0 -282 1 2.860156 1 +281 0 -4.6981187 0 +282 1 2.860157 1 286 1 12.544172 1 287 0 -4.75809 0 -289 1 6.6595697 1 +289 1 6.6595707 1 292 1 ? 0 294 0 ? 0 295 1 5.621522 1 -298 0 -2.4584546 0 -302 1 12.725584 1 +298 0 -2.4584541 0 +302 1 12.72558 1 305 1 8.040865 1 306 0 -5.4043503 0 307 0 -5.4043503 0 @@ -512,34 +512,34 @@ Instance Label Score Assigned 322 0 -4.4382315 0 324 0 -5.4043503 0 325 0 -3.7860875 0 -326 1 3.6223288 1 -330 1 4.9927454 1 +326 1 3.6223297 1 +330 1 4.9927444 1 334 1 5.514736 1 335 0 -5.935418 0 337 0 -5.4043503 0 339 1 5.1161976 1 -340 1 5.5803347 1 +340 1 5.5803356 1 341 0 -5.4043503 0 342 0 -5.4523587 0 345 0 -5.935418 0 -351 0 -4.9212914 0 +351 0 -4.921291 0 356 1 -0.9127703 0 357 1 9.874601 1 359 1 4.653248 1 362 0 -3.5174994 0 363 0 -1.6840982 0 -364 0 -4.9212914 0 +364 0 -4.921291 0 365 0 -5.186825 0 367 1 9.040358 1 -369 0 -5.1140842 0 +369 0 -5.1140847 0 372 0 -3.9699688 0 374 0 -4.4530277 0 375 0 -5.935418 0 380 0 -5.935418 0 382 0 -3.5015903 0 -385 0 -3.4273872 0 +385 0 -3.4273868 0 386 1 5.084236 1 -390 0 -5.4467115 0 +390 0 -5.446712 0 393 0 -5.881093 0 394 0 -4.909887 0 397 0 -4.48624 0 @@ -549,12 +549,12 @@ Instance Label Score Assigned 403 0 -3.7908158 0 405 0 -5.669884 0 407 0 -5.669884 0 -408 0 -3.5375085 0 +408 0 -3.537508 0 410 0 -5.669884 0 411 0 ? 0 -412 1 7.6394253 1 +412 1 7.639426 1 417 0 -5.669884 0 -420 0 -2.6969714 0 +420 0 -2.696971 0 421 1 9.498289 1 424 0 -4.9692993 0 425 1 11.849485 1 @@ -565,9 +565,9 @@ Instance Label Score Assigned 433 0 -4.013695 0 435 1 7.0216722 1 437 0 -4.48624 0 -438 0 -3.5384212 0 +438 0 -3.5384207 0 443 0 -5.54259 0 -444 0 -2.6761093 0 +444 0 -2.6761103 0 445 0 -5.4523587 0 446 0 -5.935418 0 447 0 -4.0031805 0 @@ -576,31 +576,31 @@ Instance Label Score Assigned 459 0 -3.5017061 0 460 0 -3.5568361 0 461 0 -3.3979883 0 -462 0 -3.0737762 0 +462 0 -3.0737767 0 463 0 -4.282031 0 468 0 -4.48624 0 469 0 -5.4191465 0 470 0 -4.649441 0 -471 0 -3.0737762 0 +471 0 -3.0737767 0 472 0 -3.4657965 0 473 0 -4.48624 0 475 0 -4.9692993 0 -476 0 -4.2355022 0 +476 0 -4.2355027 0 477 0 -4.48624 0 478 0 -3.744658 0 479 1 6.673233 1 481 0 -2.5658813 0 485 0 -4.6490927 0 486 0 -4.649441 0 -488 1 0.9591036 1 +488 1 0.95910263 1 490 0 -5.935418 0 -491 1 5.556223 1 +491 1 5.556222 1 494 0 -0.015696526 0 496 0 -5.881093 0 498 0 -4.220706 0 499 0 -4.220706 0 -500 0 -2.9890537 0 -503 0 -3.9551725 0 +500 0 -2.9890532 0 +503 0 -3.955172 0 505 0 -4.4522543 0 506 1 8.766861 1 508 0 -4.0031805 0 @@ -609,12 +609,12 @@ Instance Label Score Assigned 512 0 -4.0031805 0 515 1 7.3749876 1 516 0 -5.881093 0 -518 0 -4.49539 0 +518 0 -4.4953895 0 524 0 -4.7037654 0 525 0 -4.746127 0 526 0 -4.48624 0 -530 1 5.325206 1 -536 0 -3.4721131 0 +530 1 5.325205 1 +536 0 -3.4721127 0 537 0 -3.2213755 0 542 0 -3.4861355 0 543 0 -4.220706 0 @@ -622,7 +622,7 @@ Instance Label Score Assigned 550 0 -4.7037654 0 551 0 -5.4043503 0 552 0 -3.2360563 0 -553 0 -1.4225526 0 +553 0 -1.4225531 0 554 0 -4.9692993 0 555 0 -1.7431297 0 556 0 -2.9508896 0 @@ -631,7 +631,7 @@ Instance Label Score Assigned 567 0 -3.4350505 0 568 1 3.5748348 1 570 1 6.466878 1 -571 1 9.048693 1 +571 1 9.048691 1 572 0 -4.7037654 0 573 0 -5.669884 0 574 1 5.533702 1 @@ -644,15 +644,15 @@ Instance Label Score Assigned 587 0 -3.5334377 0 588 1 4.6442137 1 589 0 -4.0031805 0 -591 1 4.243066 1 -592 1 4.8517914 1 +591 1 4.243067 1 +592 1 4.8517895 1 595 0 -3.7376466 0 596 0 -3.9699688 0 597 0 -2.9706383 0 598 0 -4.7037654 0 599 0 -2.9381208 0 601 0 -5.6155596 0 -603 1 3.1762495 1 +603 1 3.1762505 1 605 1 8.159748 1 608 1 8.079367 1 610 1 6.972576 1 @@ -667,15 +667,15 @@ Instance Label Score Assigned 630 0 -2.7601237 0 631 0 -3.7376466 0 632 0 -5.935418 0 -635 0 -4.217087 0 -636 1 8.162584 1 +635 0 -4.2170873 0 +636 1 8.162586 1 637 0 -2.370799 0 -640 0 -4.0398955 0 +640 0 -4.039895 0 643 0 -5.935418 0 646 0 -5.426158 0 647 0 -5.4890733 0 -648 1 8.579458 1 -650 0 -3.6219687 0 +648 1 8.579456 1 +650 0 -3.6219683 0 651 0 -4.965017 0 655 0 -4.7037654 0 658 1 7.546403 1 @@ -683,11 +683,11 @@ Instance Label Score Assigned 662 0 -5.1684093 0 663 0 -5.1684093 0 664 0 -4.2574205 0 -666 0 -3.070212 0 +666 0 -3.0702114 0 667 0 -4.4382315 0 -669 1 6.981786 1 +669 1 6.9817867 1 671 0 -3.9565368 0 -672 0 -4.9212914 0 +672 0 -4.921291 0 673 0 -3.289723 0 674 0 -5.669884 0 675 0 -3.7340279 0 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-out.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-out.txt index 08ac6f6825..0505f9c408 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-out.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-out.txt @@ -9,28 +9,28 @@ Confusion table ||====================== PREDICTED || positive | negative | Recall TRUTH ||====================== - positive || 234 | 5 | 0.9791 + positive || 233 | 6 | 0.9749 negative || 12 | 432 | 0.9730 ||====================== -Precision || 0.9512 | 0.9886 | -OVERALL 0/1 ACCURACY: 0.975110 -LOG LOSS/instance: 0.084507 +Precision || 0.9510 | 0.9863 | +OVERALL 0/1 ACCURACY: 0.973646 +LOG LOSS/instance: 0.083944 Test-set entropy (prior Log-Loss/instance): 0.934003 -LOG-LOSS REDUCTION (RIG): 0.909522 +LOG-LOSS REDUCTION (RIG): 0.910125 AUC: 0.996146 OVERALL RESULTS --------------------------------------- AUC: 0.996146 (0.0000) -Accuracy: 0.975110 (0.0000) -Positive precision: 0.951220 (0.0000) -Positive recall: 0.979079 (0.0000) -Negative precision: 0.988558 (0.0000) +Accuracy: 0.973646 (0.0000) +Positive precision: 0.951020 (0.0000) +Positive recall: 0.974895 (0.0000) +Negative precision: 0.986301 (0.0000) Negative recall: 0.972973 (0.0000) -Log-loss: 0.084507 (0.0000) -Log-loss reduction: 0.909522 (0.0000) -F1 Score: 0.964948 (0.0000) -AUPRC: 0.992065 (0.0000) +Log-loss: 0.083944 (0.0000) +Log-loss reduction: 0.910125 (0.0000) +F1 Score: 0.962810 (0.0000) +AUPRC: 0.992010 (0.0000) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-rp.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-rp.txt index e852d92116..1b1e71c268 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-rp.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-rp.txt @@ -1,4 +1,4 @@ AveragedPerceptron AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings -0.996146 0.97511 0.95122 0.979079 0.988558 0.972973 0.084507 0.909522 0.964948 0.992065 AveragedPerceptron %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=AveragedPerceptron cali=PAV dout=%Output% data=%Data% out=%Output% seed=1 +0.996146 0.973646 0.95102 0.974895 0.986301 0.972973 0.083944 0.910125 0.96281 0.99201 AveragedPerceptron %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=AveragedPerceptron cali=PAV dout=%Output% data=%Data% out=%Output% seed=1 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration.txt index 67338259ca..651d20da42 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration.txt @@ -1,700 +1,700 @@ Instance Label Score Probability Log-loss Assigned -0 0 -3.5726995 1E-15 1.4415419267167138E-15 0 -1 0 3.6101465 0.8333333 2.5849623287385155 1 -2 0 -4.070944 1E-15 1.4415419267167138E-15 0 -3 0 2.470542 0.8095238 2.3923175087700885 1 -4 0 -3.4358397 1E-15 1.4415419267167138E-15 0 -5 1 12.382595 1 -0 1 -6 0 -1.4209604 0.071428575 0.10691520887754996 0 -7 0 -4.701088 1E-15 1.4415419267167138E-15 0 -8 0 -4.6745405 1E-15 1.4415419267167138E-15 0 -9 0 -4.406417 1E-15 1.4415419267167138E-15 0 -10 0 -5.559344 1E-15 1.4415419267167138E-15 0 -11 0 -5.4818344 1E-15 1.4415419267167138E-15 0 -12 1 -0.14206886 0.6363636 0.65207672114864346 0 -13 0 -4.5691886 1E-15 1.4415419267167138E-15 0 -14 1 9.321611 1 -0 1 -15 1 1.3856993 0.8095238 0.30485456129516797 1 -16 0 -4.533843 1E-15 1.4415419267167138E-15 0 -17 0 -4.046695 1E-15 1.4415419267167138E-15 0 -18 1 7.8903713 1 -0 1 -19 0 -3.0987039 1E-15 1.4415419267167138E-15 0 -20 1 7.528511 1 -0 1 -21 1 7.875206 1 -0 1 -22 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +0 0 -3.7050557 1E-15 1.4415419267167138E-15 0 +1 0 3.4661026 0.85714287 2.807355008048932 1 +2 0 -4.24041 1E-15 1.4415419267167138E-15 0 +3 0 2.1493654 0.8 2.3219281808786905 1 +4 0 -3.5576057 1E-15 1.4415419267167138E-15 0 +5 1 12.155442 1 -0 1 +6 0 -1.573504 0.083333336 0.12553088599255555 0 +7 0 -4.9137383 1E-15 1.4415419267167138E-15 0 +8 0 -4.903472 1E-15 1.4415419267167138E-15 0 +9 0 -4.57476 1E-15 1.4415419267167138E-15 0 +10 0 -5.7696943 1E-15 1.4415419267167138E-15 0 +11 0 -5.679842 1E-15 1.4415419267167138E-15 0 +12 1 -0.2966156 0.6923077 0.53051467848041345 0 +13 0 -4.7757645 1E-15 1.4415419267167138E-15 0 +14 1 9.220881 1 -0 1 +15 1 1.1802778 0.8 0.32192807338953117 1 +16 0 -4.687022 1E-15 1.4415419267167138E-15 0 +17 0 -4.2014656 1E-15 1.4415419267167138E-15 0 +18 1 7.81973 1 -0 1 +19 0 -3.2086458 1E-15 1.4415419267167138E-15 0 +20 1 7.5404253 1 -0 1 +21 1 7.730527 1 -0 1 +22 0 -5.183432 1E-15 1.4415419267167138E-15 0 23 1 ? ? ? 0 -24 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -25 1 1.741828 0.8095238 0.30485456129516797 1 -26 0 -4.9710746 1E-15 1.4415419267167138E-15 0 -27 0 -4.0598474 1E-15 1.4415419267167138E-15 0 -28 0 -5.4818344 1E-15 1.4415419267167138E-15 0 -29 0 -5.8557806 1E-15 1.4415419267167138E-15 0 -30 0 -5.0985007 1E-15 1.4415419267167138E-15 0 -31 0 -4.9946866 1E-15 1.4415419267167138E-15 0 -32 1 7.46414 1 -0 1 -33 0 -4.689259 1E-15 1.4415419267167138E-15 0 -34 0 -4.71424 1E-15 1.4415419267167138E-15 0 -35 0 -5.4818344 1E-15 1.4415419267167138E-15 0 -36 1 9.099108 1 -0 1 -37 0 -1.113348 0.071428575 0.10691520887754996 0 -38 1 6.140953 0.98 0.029146317580716615 1 -39 1 2.5109034 0.8095238 0.30485456129516797 1 +24 0 -5.690696 1E-15 1.4415419267167138E-15 0 +25 1 1.6430082 0.8 0.32192807338953117 1 +26 0 -5.1501684 1E-15 1.4415419267167138E-15 0 +27 0 -4.190612 1E-15 1.4415419267167138E-15 0 +28 0 -5.679842 1E-15 1.4415419267167138E-15 0 +29 0 -6.1007133 1E-15 1.4415419267167138E-15 0 +30 0 -5.2624307 1E-15 1.4415419267167138E-15 0 +31 0 -5.194286 1E-15 1.4415419267167138E-15 0 +32 1 7.2332277 1 -0 1 +33 0 -4.872356 1E-15 1.4415419267167138E-15 0 +34 0 -4.9028845 1E-15 1.4415419267167138E-15 0 +35 0 -5.679842 1E-15 1.4415419267167138E-15 0 +36 1 8.852162 1 -0 1 +37 0 -1.2331572 0.083333336 0.12553088599255555 0 +38 1 6.101239 0.9814815 0.026967031375491075 1 +39 1 2.2340298 0.8 0.32192807338953117 1 40 0 ? ? ? 0 -41 1 3.3300762 0.8333333 0.26303444023032446 1 -42 1 8.577511 1 -0 1 -43 1 0.49126053 0.6363636 0.65207672114864346 1 -44 1 8.255751 1 -0 1 -45 0 -5.6322193 1E-15 1.4415419267167138E-15 0 -46 1 4.5673857 0.9285714 0.10691524360481655 1 -47 0 -5.95583 1E-15 1.4415419267167138E-15 0 -48 0 -3.4358397 1E-15 1.4415419267167138E-15 0 -49 1 5.3666544 0.98 0.029146317580716615 1 -50 1 2.5949678 0.8095238 0.30485456129516797 1 -51 1 0.12595749 0.6363636 0.65207672114864346 1 -52 1 5.2992115 0.98 0.029146317580716615 1 -53 1 8.407228 1 -0 1 -54 1 7.649311 1 -0 1 -55 1 4.478709 0.9285714 0.10691524360481655 1 -56 1 5.5541325 0.98 0.029146317580716615 1 -57 1 1.6657066 0.8095238 0.30485456129516797 1 -58 1 2.5265894 0.8095238 0.30485456129516797 1 -59 1 1.7368536 0.8095238 0.30485456129516797 1 -60 1 2.3288136 0.8095238 0.30485456129516797 1 -61 0 -5.5060835 1E-15 1.4415419267167138E-15 0 -62 1 6.380088 0.98 0.029146317580716615 1 -63 1 0.3348999 0.6363636 0.65207672114864346 1 -64 0 -5.95583 1E-15 1.4415419267167138E-15 0 -65 1 3.8072634 0.9285714 0.10691524360481655 1 -66 0 -4.046695 1E-15 1.4415419267167138E-15 0 -67 1 4.218014 0.9285714 0.10691524360481655 1 -68 1 10.826723 1 -0 1 -69 0 -5.2716546 1E-15 1.4415419267167138E-15 0 -70 0 -3.4726496 1E-15 1.4415419267167138E-15 0 -71 1 7.895048 1 -0 1 -72 0 -2.1755848 0.071428575 0.10691520887754996 0 -73 1 8.9055195 1 -0 1 -74 1 2.5993576 0.8095238 0.30485456129516797 1 -75 0 -4.0411606 1E-15 1.4415419267167138E-15 0 -76 0 -5.075033 1E-15 1.4415419267167138E-15 0 -77 0 -3.4995675 1E-15 1.4415419267167138E-15 0 -78 0 -3.6211967 1E-15 1.4415419267167138E-15 0 -79 0 -5.3911724 1E-15 1.4415419267167138E-15 0 -80 0 -2.7157316 1E-15 1.4415419267167138E-15 0 -81 0 -4.2284155 1E-15 1.4415419267167138E-15 0 -82 0 -3.4452734 1E-15 1.4415419267167138E-15 0 -83 0 -2.1223516 0.071428575 0.10691520887754996 0 -84 1 9.694054 1 -0 1 -85 1 6.2895603 0.98 0.029146317580716615 1 -86 1 2.6168842 0.8095238 0.30485456129516797 1 -87 1 6.91914 1 -0 1 -88 0 -4.046695 1E-15 1.4415419267167138E-15 0 -89 0 -5.085745 1E-15 1.4415419267167138E-15 0 -90 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -91 0 -5.189559 1E-15 1.4415419267167138E-15 0 -92 0 -4.046695 1E-15 1.4415419267167138E-15 0 -93 0 -5.95583 1E-15 1.4415419267167138E-15 0 -94 0 -4.9946866 1E-15 1.4415419267167138E-15 0 -95 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -96 0 -5.663555 1E-15 1.4415419267167138E-15 0 -97 0 -3.5726995 1E-15 1.4415419267167138E-15 0 -98 1 8.590233 1 -0 1 -99 1 10.917194 1 -0 1 -100 1 4.8476696 0.9285714 0.10691524360481655 1 -101 1 -0.842803 0.5 1 0 -102 0 -3.7530966 1E-15 1.4415419267167138E-15 0 -103 1 1.7746449 0.8095238 0.30485456129516797 1 -104 1 12.140858 1 -0 1 -105 1 2.5560703 0.8095238 0.30485456129516797 1 -106 1 9.259369 1 -0 1 -107 1 6.720646 0.98 0.029146317580716615 1 -108 0 -5.5617743 1E-15 1.4415419267167138E-15 0 -109 1 6.871727 1 -0 1 -110 0 -2.766693 1E-15 1.4415419267167138E-15 0 -111 1 3.848031 0.9285714 0.10691524360481655 1 -112 1 9.42577 1 -0 1 -113 1 9.506622 1 -0 1 -114 0 -3.0727453 1E-15 1.4415419267167138E-15 0 -115 0 -4.643991 1E-15 1.4415419267167138E-15 0 -116 0 -0.6618881 0.5 1 0 -117 1 9.617277 1 -0 1 -118 0 -5.3621607 1E-15 1.4415419267167138E-15 0 -119 0 -3.9435177 1E-15 1.4415419267167138E-15 0 -120 0 -4.8696556 1E-15 1.4415419267167138E-15 0 -121 0 -3.469522 1E-15 1.4415419267167138E-15 0 -122 1 9.680523 1 -0 1 -123 1 3.8165932 0.9285714 0.10691524360481655 1 -124 1 7.6522446 1 -0 1 -125 0 -5.95583 1E-15 1.4415419267167138E-15 0 -126 1 8.564951 1 -0 1 -127 0 -4.520691 1E-15 1.4415419267167138E-15 0 -128 1 4.848981 0.9285714 0.10691524360481655 1 -129 0 -5.717684 1E-15 1.4415419267167138E-15 0 -130 0 -3.4726496 1E-15 1.4415419267167138E-15 0 -131 0 -4.9946866 1E-15 1.4415419267167138E-15 0 -132 1 8.60223 1 -0 1 -133 0 -4.810811 1E-15 1.4415419267167138E-15 0 -134 0 -4.917177 1E-15 1.4415419267167138E-15 0 -135 0 -2.7288966 1E-15 1.4415419267167138E-15 0 -136 0 -4.533843 1E-15 1.4415419267167138E-15 0 -137 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -138 0 -4.2402444 1E-15 1.4415419267167138E-15 0 +41 1 3.2122545 0.85714287 0.22239240700456045 1 +42 1 8.198594 1 -0 1 +43 1 0.22008419 0.6923077 0.53051467848041345 1 +44 1 7.915265 1 -0 1 +45 0 -5.8607707 1E-15 1.4415419267167138E-15 0 +46 1 4.2778234 0.9230769 0.11547721025399223 1 +47 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +48 0 -3.5576057 1E-15 1.4415419267167138E-15 0 +49 1 5.1043253 0.9814815 0.026967031375491075 1 +50 1 2.3847685 0.8 0.32192807338953117 1 +51 1 -0.002776146 0.6923077 0.53051467848041345 0 +52 1 5.171505 0.9814815 0.026967031375491075 1 +53 1 8.174706 1 -0 1 +54 1 7.5763617 1 -0 1 +55 1 4.3368845 0.9230769 0.11547721025399223 1 +56 1 5.228385 0.9814815 0.026967031375491075 1 +57 1 1.4711056 0.8 0.32192807338953117 1 +58 1 2.3726006 0.8 0.32192807338953117 1 +59 1 1.6023321 0.8 0.32192807338953117 1 +60 1 2.1442842 0.8 0.32192807338953117 1 +61 0 -5.7187867 1E-15 1.4415419267167138E-15 0 +62 1 6.0357714 0.9814815 0.026967031375491075 1 +63 1 0.09457874 0.6923077 0.53051467848041345 1 +64 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +65 1 3.6413336 0.9230769 0.11547721025399223 1 +66 0 -4.2014656 1E-15 1.4415419267167138E-15 0 +67 1 3.980445 0.9230769 0.11547721025399223 1 +68 1 10.612863 1 -0 1 +69 0 -5.4978714 1E-15 1.4415419267167138E-15 0 +70 0 -3.6295166 1E-15 1.4415419267167138E-15 0 +71 1 7.5565577 1 -0 1 +72 0 -2.4480972 1E-15 1.4415419267167138E-15 0 +73 1 8.905442 1 -0 1 +74 1 2.4551315 0.8 0.32192807338953117 1 +75 0 -4.274302 1E-15 1.4415419267167138E-15 0 +76 0 -5.334609 1E-15 1.4415419267167138E-15 0 +77 0 -3.644158 1E-15 1.4415419267167138E-15 0 +78 0 -3.7829442 1E-15 1.4415419267167138E-15 0 +79 0 -5.6008434 1E-15 1.4415419267167138E-15 0 +80 0 -2.9382496 1E-15 1.4415419267167138E-15 0 +81 0 -4.3650923 1E-15 1.4415419267167138E-15 0 +82 0 -3.5927935 1E-15 1.4415419267167138E-15 0 +83 0 -2.2981849 1E-15 1.4415419267167138E-15 0 +84 1 9.3922205 1 -0 1 +85 1 5.7727385 0.9814815 0.026967031375491075 1 +86 1 2.3984623 0.8 0.32192807338953117 1 +87 1 6.6674128 0.9814815 0.026967031375491075 1 +88 0 -4.2014656 1E-15 1.4415419267167138E-15 0 +89 0 -5.2789135 1E-15 1.4415419267167138E-15 0 +90 0 -5.690696 1E-15 1.4415419267167138E-15 0 +91 0 -5.3470583 1E-15 1.4415419267167138E-15 0 +92 0 -4.2014656 1E-15 1.4415419267167138E-15 0 +93 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +94 0 -5.194286 1E-15 1.4415419267167138E-15 0 +95 0 -5.690696 1E-15 1.4415419267167138E-15 0 +96 0 -5.8434687 1E-15 1.4415419267167138E-15 0 +97 0 -3.7050557 1E-15 1.4415419267167138E-15 0 +98 1 8.223899 1 -0 1 +99 1 10.822373 1 -0 1 +100 1 4.636462 0.9230769 0.11547721025399223 1 +101 1 -1.0467243 0.4 1.3219280733895313 0 +102 0 -3.9209185 1E-15 1.4415419267167138E-15 0 +103 1 1.630683 0.8 0.32192807338953117 1 +104 1 11.827565 1 -0 1 +105 1 2.3993664 0.8 0.32192807338953117 1 +106 1 9.16854 1 -0 1 +107 1 6.4510603 0.9814815 0.026967031375491075 1 +108 0 -5.7906036 1E-15 1.4415419267167138E-15 0 +109 1 6.7095194 0.9814815 0.026967031375491075 1 +110 0 -3.014533 1E-15 1.4415419267167138E-15 0 +111 1 3.7277098 0.9230769 0.11547721025399223 1 +112 1 9.536014 1 -0 1 +113 1 9.218173 1 -0 1 +114 0 -3.277872 1E-15 1.4415419267167138E-15 0 +115 0 -4.8319454 1E-15 1.4415419267167138E-15 0 +116 0 -0.7799158 0.4 0.73696560849809378 0 +117 1 9.464227 1 -0 1 +118 0 -5.5888796 1E-15 1.4415419267167138E-15 0 +119 0 -4.1281476 1E-15 1.4415419267167138E-15 0 +120 0 -5.088218 1E-15 1.4415419267167138E-15 0 +121 0 -3.6317377 1E-15 1.4415419267167138E-15 0 +122 1 9.371118 1 -0 1 +123 1 3.5896034 0.9230769 0.11547721025399223 1 +124 1 7.4270334 1 -0 1 +125 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +126 1 8.480126 1 -0 1 +127 0 -4.697876 1E-15 1.4415419267167138E-15 0 +128 1 4.7593575 0.9230769 0.11547721025399223 1 +129 0 -6.02982 1E-15 1.4415419267167138E-15 0 +130 0 -3.6295166 1E-15 1.4415419267167138E-15 0 +131 0 -5.194286 1E-15 1.4415419267167138E-15 0 +132 1 8.352777 1 -0 1 +133 0 -4.9906077 1E-15 1.4415419267167138E-15 0 +134 0 -5.1044335 1E-15 1.4415419267167138E-15 0 +135 0 -2.8619413 1E-15 1.4415419267167138E-15 0 +136 0 -4.687022 1E-15 1.4415419267167138E-15 0 +137 0 -5.668988 1E-15 1.4415419267167138E-15 0 +138 0 -4.4064746 1E-15 1.4415419267167138E-15 0 139 0 ? ? ? 0 -140 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -141 0 -5.9689827 1E-15 1.4415419267167138E-15 0 -142 1 4.4324036 0.9285714 0.10691524360481655 1 -143 0 -4.643991 1E-15 1.4415419267167138E-15 0 -144 0 -5.4818344 1E-15 1.4415419267167138E-15 0 +140 0 -5.668988 1E-15 1.4415419267167138E-15 0 +141 0 -6.1653986 1E-15 1.4415419267167138E-15 0 +142 1 4.307641 0.9230769 0.11547721025399223 1 +143 0 -4.8319454 1E-15 1.4415419267167138E-15 0 +144 0 -5.679842 1E-15 1.4415419267167138E-15 0 145 0 ? ? ? 0 -146 1 1.3394356 0.8095238 0.30485456129516797 1 -147 0 -5.4154215 1E-15 1.4415419267167138E-15 0 -148 0 -1.012373 0.071428575 0.10691520887754996 0 -149 1 11.461615 1 -0 1 -150 0 -5.559344 1E-15 1.4415419267167138E-15 0 -151 1 5.006485 0.9285714 0.10691524360481655 1 -152 1 9.715748 1 -0 1 -153 0 -4.1214976 1E-15 1.4415419267167138E-15 0 -154 0 -6.442978 1E-15 1.4415419267167138E-15 0 -155 1 3.7769232 0.9285714 0.10691524360481655 1 -156 0 -5.5348053 1E-15 1.4415419267167138E-15 0 -157 0 -4.9946866 1E-15 1.4415419267167138E-15 0 +146 1 1.0868425 0.6923077 0.53051467848041345 1 +147 0 -5.639788 1E-15 1.4415419267167138E-15 0 +148 0 -1.259141 0.083333336 0.12553088599255555 0 +149 1 11.153153 1 -0 1 +150 0 -5.7696943 1E-15 1.4415419267167138E-15 0 +151 1 4.855405 0.9814815 0.026967031375491075 1 +152 1 9.511897 1 -0 1 +153 0 -4.257647 1E-15 1.4415419267167138E-15 0 +154 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +155 1 3.6720238 0.9230769 0.11547721025399223 1 +156 0 -5.7834425 1E-15 1.4415419267167138E-15 0 +157 0 -5.194286 1E-15 1.4415419267167138E-15 0 158 0 ? ? ? 0 -159 1 12.346203 1 -0 1 -160 1 9.039494 1 -0 1 -161 0 -3.8496675 1E-15 1.4415419267167138E-15 0 -162 0 -4.520691 1E-15 1.4415419267167138E-15 0 -163 0 -3.387055 1E-15 1.4415419267167138E-15 0 +159 1 12.03571 1 -0 1 +160 1 8.89737 1 -0 1 +161 0 -4.0086412 1E-15 1.4415419267167138E-15 0 +162 0 -4.697876 1E-15 1.4415419267167138E-15 0 +163 0 -3.6110554 1E-15 1.4415419267167138E-15 0 164 0 ? ? ? 0 -165 0 -3.3999205 1E-15 1.4415419267167138E-15 0 -166 1 7.976183 1 -0 1 -167 1 8.355644 1 -0 1 -168 0 -4.520691 1E-15 1.4415419267167138E-15 0 -169 0 -6.2282124 1E-15 1.4415419267167138E-15 0 -170 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -171 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -172 0 -5.95583 1E-15 1.4415419267167138E-15 0 -173 1 15.1560135 1 -0 1 -174 1 6.1769247 0.98 0.029146317580716615 1 -175 1 7.842922 1 -0 1 -176 0 -4.9946866 1E-15 1.4415419267167138E-15 0 -177 1 4.766121 0.9285714 0.10691524360481655 1 -178 0 -4.046695 1E-15 1.4415419267167138E-15 0 -179 1 2.290575 0.8095238 0.30485456129516797 1 -180 0 -5.559344 1E-15 1.4415419267167138E-15 0 -181 0 -6.442978 1E-15 1.4415419267167138E-15 0 -182 0 -3.0987039 1E-15 1.4415419267167138E-15 0 -183 1 9.159964 1 -0 1 -184 1 6.2014647 0.98 0.029146317580716615 1 -185 0 -5.0853486 1E-15 1.4415419267167138E-15 0 -186 1 5.7654104 0.98 0.029146317580716615 1 -187 1 13.977451 1 -0 1 -188 1 9.065283 1 -0 1 -189 0 -4.7540584 1E-15 1.4415419267167138E-15 0 -190 1 11.957218 1 -0 1 -191 1 10.956873 1 -0 1 -192 0 -4.0598474 1E-15 1.4415419267167138E-15 0 -193 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -194 0 -4.520691 1E-15 1.4415419267167138E-15 0 -195 0 -4.046695 1E-15 1.4415419267167138E-15 0 -196 0 6.8652763 0.98 5.6438575656365879 1 -197 0 -2.6564164 1E-15 1.4415419267167138E-15 0 -198 0 -6.442978 1E-15 1.4415419267167138E-15 0 -199 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -200 1 10.36586 1 -0 1 -201 1 9.869495 1 -0 1 -202 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -203 0 -3.5726995 1E-15 1.4415419267167138E-15 0 -204 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -205 1 12.086601 1 -0 1 -206 1 5.944169 0.98 0.029146317580716615 1 -207 0 -5.559344 1E-15 1.4415419267167138E-15 0 -208 0 -5.559344 1E-15 1.4415419267167138E-15 0 -209 0 -3.6633615 1E-15 1.4415419267167138E-15 0 -210 1 14.534113 1 -0 1 -211 1 9.64962 1 -0 1 -212 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -213 1 14.529058 1 -0 1 -214 1 13.868914 1 -0 1 -215 1 7.643734 1 -0 1 -216 0 -5.95583 1E-15 1.4415419267167138E-15 0 -217 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -218 1 7.8867817 1 -0 1 -219 0 -2.511506 1E-15 1.4415419267167138E-15 0 -220 0 -5.1632547 1E-15 1.4415419267167138E-15 0 -221 1 10.395218 1 -0 1 -222 1 -2.214662 0.071428575 3.8073548575641118 0 -223 1 5.7424126 0.98 0.029146317580716615 1 -224 1 9.995327 1 -0 1 -225 0 -5.95583 1E-15 1.4415419267167138E-15 0 -226 1 10.225868 1 -0 1 -227 1 7.459608 1 -0 1 -228 0 -5.559344 1E-15 1.4415419267167138E-15 0 -229 1 12.666515 1 -0 1 -230 1 6.1583214 0.98 0.029146317580716615 1 -231 1 8.623034 1 -0 1 -232 0 1.2822819 0.6363636 1.4594315756416352 1 -233 1 6.3825197 0.98 0.029146317580716615 1 -234 0 -2.8964381 1E-15 1.4415419267167138E-15 0 +165 0 -3.5511756 1E-15 1.4415419267167138E-15 0 +166 1 7.750617 1 -0 1 +167 1 8.140458 1 -0 1 +168 0 -4.697876 1E-15 1.4415419267167138E-15 0 +169 0 -6.418877 1E-15 1.4415419267167138E-15 0 +170 0 -5.668988 1E-15 1.4415419267167138E-15 0 +171 0 -5.690696 1E-15 1.4415419267167138E-15 0 +172 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +173 1 14.842399 1 -0 1 +174 1 6.0863256 0.9814815 0.026967031375491075 1 +175 1 7.5728846 1 -0 1 +176 0 -5.194286 1E-15 1.4415419267167138E-15 0 +177 1 4.2855787 0.9230769 0.11547721025399223 1 +178 0 -4.2014656 1E-15 1.4415419267167138E-15 0 +179 1 2.15761 0.8 0.32192807338953117 1 +180 0 -5.7696943 1E-15 1.4415419267167138E-15 0 +181 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +182 0 -3.2086458 1E-15 1.4415419267167138E-15 0 +183 1 8.944306 1 -0 1 +184 1 6.0725775 0.9814815 0.026967031375491075 1 +185 0 -5.2732844 1E-15 1.4415419267167138E-15 0 +186 1 5.448736 0.9814815 0.026967031375491075 1 +187 1 13.897843 1 -0 1 +188 1 8.907805 1 -0 1 +189 0 -5.0173383 1E-15 1.4415419267167138E-15 0 +190 1 11.761691 1 -0 1 +191 1 10.76023 1 -0 1 +192 0 -4.190612 1E-15 1.4415419267167138E-15 0 +193 0 -5.690696 1E-15 1.4415419267167138E-15 0 +194 0 -4.697876 1E-15 1.4415419267167138E-15 0 +195 0 -4.2014656 1E-15 1.4415419267167138E-15 0 +196 0 6.7605906 0.9814815 5.7548883620769793 1 +197 0 -2.7392664 1E-15 1.4415419267167138E-15 0 +198 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +199 0 -5.183432 1E-15 1.4415419267167138E-15 0 +200 1 10.20231 1 -0 1 +201 1 9.669415 1 -0 1 +202 0 -5.690696 1E-15 1.4415419267167138E-15 0 +203 0 -3.7050557 1E-15 1.4415419267167138E-15 0 +204 0 -5.690696 1E-15 1.4415419267167138E-15 0 +205 1 11.778428 1 -0 1 +206 1 5.671749 0.9814815 0.026967031375491075 1 +207 0 -5.7696943 1E-15 1.4415419267167138E-15 0 +208 0 -5.7696943 1E-15 1.4415419267167138E-15 0 +209 0 -3.7840543 1E-15 1.4415419267167138E-15 0 +210 1 14.311627 1 -0 1 +211 1 9.471307 1 -0 1 +212 0 -5.690696 1E-15 1.4415419267167138E-15 0 +213 1 14.345925 1 -0 1 +214 1 13.768859 1 -0 1 +215 1 7.442033 1 -0 1 +216 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +217 0 -5.690696 1E-15 1.4415419267167138E-15 0 +218 1 7.5836535 1 -0 1 +219 0 -2.6475506 1E-15 1.4415419267167138E-15 0 +220 0 -5.368766 1E-15 1.4415419267167138E-15 0 +221 1 10.2217455 1 -0 1 +222 1 -2.2922978 0.083333336 3.5849624577254944 0 +223 1 5.585368 0.9814815 0.026967031375491075 1 +224 1 9.954375 1 -0 1 +225 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +226 1 10.120683 1 -0 1 +227 1 7.2224054 1 -0 1 +228 0 -5.7696943 1E-15 1.4415419267167138E-15 0 +229 1 12.354081 1 -0 1 +230 1 6.003766 0.9814815 0.026967031375491075 1 +231 1 8.352307 1 -0 1 +232 0 1.1177626 0.6923077 1.7004398041324202 1 +233 1 6.363435 0.9814815 0.026967031375491075 1 +234 0 -3.1496835 1E-15 1.4415419267167138E-15 0 235 0 ? ? ? 0 -236 1 11.420414 1 -0 1 -237 1 6.535795 0.98 0.029146317580716615 1 -238 1 12.422876 1 -0 1 -239 1 5.9025297 0.98 0.029146317580716615 1 -240 0 -2.0179915 0.071428575 0.10691520887754996 0 -241 0 -4.0004973 1E-15 1.4415419267167138E-15 0 -242 0 -4.9946866 1E-15 1.4415419267167138E-15 0 -243 0 -2.6953988 1E-15 1.4415419267167138E-15 0 -244 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -245 0 -2.817525 1E-15 1.4415419267167138E-15 0 -246 1 11.424002 1 -0 1 -247 1 3.104393 0.8333333 0.26303444023032446 1 -248 0 -3.0615559 1E-15 1.4415419267167138E-15 0 +236 1 11.114438 1 -0 1 +237 1 6.258853 0.9814815 0.026967031375491075 1 +238 1 12.123448 1 -0 1 +239 1 5.86524 0.9814815 0.026967031375491075 1 +240 0 -2.1425428 0.083333336 0.12553088599255555 0 +241 0 -4.1330147 1E-15 1.4415419267167138E-15 0 +242 0 -5.194286 1E-15 1.4415419267167138E-15 0 +243 0 -2.8897204 1E-15 1.4415419267167138E-15 0 +244 0 -5.690696 1E-15 1.4415419267167138E-15 0 +245 0 -2.9256601 1E-15 1.4415419267167138E-15 0 +246 1 11.219155 1 -0 1 +247 1 3.0079794 0.85714287 0.22239240700456045 1 +248 0 -3.2373443 1E-15 1.4415419267167138E-15 0 249 0 ? ? ? 0 -250 0 -6.021953 1E-15 1.4415419267167138E-15 0 -251 1 8.872498 1 -0 1 -252 0 4.5387735 0.9285714 3.8073544061097437 1 -253 1 8.577511 1 -0 1 -254 1 6.380088 0.98 0.029146317580716615 1 -255 1 4.052039 0.9285714 0.10691524360481655 1 -256 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -257 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -258 0 -4.520691 1E-15 1.4415419267167138E-15 0 -259 0 2.9647484 0.8095238 2.3923175087700885 1 -260 1 9.870926 1 -0 1 -261 1 12.206299 1 -0 1 -262 1 9.653839 1 -0 1 -263 1 8.981979 1 -0 1 -264 1 5.664708 0.98 0.029146317580716615 1 -265 0 -2.4948754 1E-15 1.4415419267167138E-15 0 -266 1 7.3661633 1 -0 1 -267 1 3.3009605 0.8333333 0.26303444023032446 1 -268 1 9.372967 1 -0 1 -269 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -270 1 6.031377 0.98 0.029146317580716615 1 -271 0 -3.5726995 1E-15 1.4415419267167138E-15 0 -272 1 3.3009605 0.8333333 0.26303444023032446 1 -273 1 0.21747208 0.6363636 0.65207672114864346 1 -274 0 -4.323663 1E-15 1.4415419267167138E-15 0 +250 0 -6.268999 1E-15 1.4415419267167138E-15 0 +251 1 8.957709 1 -0 1 +252 0 4.4631453 0.9230769 3.7004398041324205 1 +253 1 8.198594 1 -0 1 +254 1 6.0357714 0.9814815 0.026967031375491075 1 +255 1 3.8202343 0.9230769 0.11547721025399223 1 +256 0 -5.668988 1E-15 1.4415419267167138E-15 0 +257 0 -5.183432 1E-15 1.4415419267167138E-15 0 +258 0 -4.697876 1E-15 1.4415419267167138E-15 0 +259 0 2.6695328 0.8 2.3219281808786905 1 +260 1 9.885512 1 -0 1 +261 1 11.814006 1 -0 1 +262 1 9.426362 1 -0 1 +263 1 9.02839 1 -0 1 +264 1 5.4595184 0.9814815 0.026967031375491075 1 +265 0 -2.6705885 1E-15 1.4415419267167138E-15 0 +266 1 7.2168283 1 -0 1 +267 1 3.1091843 0.85714287 0.22239240700456045 1 +268 1 9.093087 1 -0 1 +269 0 -5.690696 1E-15 1.4415419267167138E-15 0 +270 1 5.858514 0.9814815 0.026967031375491075 1 +271 0 -3.7050557 1E-15 1.4415419267167138E-15 0 +272 1 3.1091843 0.85714287 0.22239240700456045 1 +273 1 0.07871342 0.6923077 0.53051467848041345 1 +274 0 -4.5050516 1E-15 1.4415419267167138E-15 0 275 0 ? ? ? 0 -276 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -277 0 -5.95583 1E-15 1.4415419267167138E-15 0 -278 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -279 1 7.127905 1 -0 1 -280 0 -4.520691 1E-15 1.4415419267167138E-15 0 -281 0 -4.689259 1E-15 1.4415419267167138E-15 0 -282 1 4.4381237 0.9285714 0.10691524360481655 1 -283 1 6.0636253 0.98 0.029146317580716615 1 -284 1 7.431343 1 -0 1 -285 1 14.218481 1 -0 1 -286 1 15.281263 1 -0 1 -287 0 -4.917177 1E-15 1.4415419267167138E-15 0 -288 1 2.2163515 0.8095238 0.30485456129516797 1 -289 1 8.312019 1 -0 1 -290 0 -6.442978 1E-15 1.4415419267167138E-15 0 -291 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +276 0 -5.183432 1E-15 1.4415419267167138E-15 0 +277 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +278 0 -5.690696 1E-15 1.4415419267167138E-15 0 +279 1 6.9710016 1 -0 1 +280 0 -4.697876 1E-15 1.4415419267167138E-15 0 +281 0 -4.872356 1E-15 1.4415419267167138E-15 0 +282 1 4.3149786 0.9230769 0.11547721025399223 1 +283 1 5.990473 0.9814815 0.026967031375491075 1 +284 1 7.4265547 1 -0 1 +285 1 13.994821 1 -0 1 +286 1 15.019143 1 -0 1 +287 0 -5.1044335 1E-15 1.4415419267167138E-15 0 +288 1 2.0468407 0.8 0.32192807338953117 1 +289 1 8.113109 1 -0 1 +290 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +291 0 -5.690696 1E-15 1.4415419267167138E-15 0 292 1 ? ? ? 0 -293 1 5.542122 0.98 0.029146317580716615 1 +293 1 5.5823574 0.9814815 0.026967031375491075 1 294 0 ? ? ? 0 -295 1 7.7866364 1 -0 1 -296 0 1.823431 0.8095238 2.3923175087700885 1 +295 1 7.634673 1 -0 1 +296 0 1.6353331 0.8 2.3219281808786905 1 297 0 ? ? ? 0 -298 0 -2.725597 1E-15 1.4415419267167138E-15 0 -299 1 7.8274364 1 -0 1 -300 1 7.348074 1 -0 1 -301 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -302 1 15.735764 1 -0 1 -303 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -304 1 5.9607973 0.98 0.029146317580716615 1 -305 1 8.459469 1 -0 1 -306 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -307 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -308 1 7.4225903 1 -0 1 -309 0 -1.7474074 0.071428575 0.10691520887754996 0 -310 0 -5.3911724 1E-15 1.4415419267167138E-15 0 -311 0 -6.442978 1E-15 1.4415419267167138E-15 0 -312 1 3.6294699 0.9285714 0.10691524360481655 1 -313 0 -6.442978 1E-15 1.4415419267167138E-15 0 -314 0 -6.0464916 1E-15 1.4415419267167138E-15 0 +298 0 -2.8376803 1E-15 1.4415419267167138E-15 0 +299 1 7.8606234 1 -0 1 +300 1 7.1416483 1 -0 1 +301 0 -5.690696 1E-15 1.4415419267167138E-15 0 +302 1 15.431024 1 -0 1 +303 0 -5.690696 1E-15 1.4415419267167138E-15 0 +304 1 6.007621 0.9814815 0.026967031375491075 1 +305 1 8.327315 1 -0 1 +306 0 -5.690696 1E-15 1.4415419267167138E-15 0 +307 0 -5.690696 1E-15 1.4415419267167138E-15 0 +308 1 7.1634607 1 -0 1 +309 0 -1.8969002 0.083333336 0.12553088599255555 0 +310 0 -5.6008434 1E-15 1.4415419267167138E-15 0 +311 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +312 1 3.4877834 0.9230769 0.11547721025399223 1 +313 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +314 0 -6.255251 1E-15 1.4415419267167138E-15 0 315 0 ? ? ? 0 -316 1 3.6177397 0.9285714 0.10691524360481655 1 -317 1 9.215706 1 -0 1 -318 0 -5.1966968 1E-15 1.4415419267167138E-15 0 -319 0 2.6369457 0.8095238 2.3923175087700885 1 -320 1 7.3824844 1 -0 1 +316 1 3.4135065 0.85714287 0.22239240700456045 1 +317 1 9.002616 1 -0 1 +318 0 -5.4537 1E-15 1.4415419267167138E-15 0 +319 0 2.4164848 0.8 2.3219281808786905 1 +320 1 7.184394 1 -0 1 321 0 ? ? ? 0 -322 0 -4.520691 1E-15 1.4415419267167138E-15 0 -323 1 5.5612926 0.98 0.029146317580716615 1 -324 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -325 0 -4.1927576 1E-15 1.4415419267167138E-15 0 -326 1 4.4103804 0.9285714 0.10691524360481655 1 -327 0 -5.95583 1E-15 1.4415419267167138E-15 0 -328 1 3.373887 0.8333333 0.26303444023032446 1 -329 1 7.8321342 1 -0 1 -330 1 5.8562517 0.98 0.029146317580716615 1 -331 0 -3.2490892 1E-15 1.4415419267167138E-15 0 -332 0 -3.1363668 1E-15 1.4415419267167138E-15 0 -333 1 4.914962 0.9285714 0.10691524360481655 1 -334 1 5.9119453 0.98 0.029146317580716615 1 -335 0 -6.442978 1E-15 1.4415419267167138E-15 0 -336 1 5.54352 0.98 0.029146317580716615 1 -337 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -338 0 -6.0464916 1E-15 1.4415419267167138E-15 0 -339 1 5.684024 0.98 0.029146317580716615 1 -340 1 6.620782 0.98 0.029146317580716615 1 -341 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -342 0 -5.9689827 1E-15 1.4415419267167138E-15 0 -343 0 -6.442978 1E-15 1.4415419267167138E-15 0 -344 1 10.162451 1 -0 1 -345 0 -6.442978 1E-15 1.4415419267167138E-15 0 -346 0 -3.335825 1E-15 1.4415419267167138E-15 0 -347 0 -6.139584 1E-15 1.4415419267167138E-15 0 -348 1 0.15727425 0.6363636 0.65207672114864346 1 -349 1 4.0622606 0.9285714 0.10691524360481655 1 -350 0 -3.93614 1E-15 1.4415419267167138E-15 0 -351 0 -4.9946866 1E-15 1.4415419267167138E-15 0 -352 0 0.4719286 0.6363636 1.4594315756416352 1 -353 1 8.696344 1 -0 1 -354 0 -5.95583 1E-15 1.4415419267167138E-15 0 -355 0 -4.246154 1E-15 1.4415419267167138E-15 0 -356 1 -0.69921684 0.5 1 0 -357 1 12.852016 1 -0 1 -358 1 5.5822067 0.98 0.029146317580716615 1 -359 1 5.3672857 0.98 0.029146317580716615 1 -360 1 15.333874 1 -0 1 -361 1 6.31769 0.98 0.029146317580716615 1 -362 0 -3.5059962 1E-15 1.4415419267167138E-15 0 -363 0 -2.065846 0.071428575 0.10691520887754996 0 -364 0 -4.9946866 1E-15 1.4415419267167138E-15 0 -365 0 -5.4818344 1E-15 1.4415419267167138E-15 0 -366 1 13.694569 1 -0 1 -367 1 11.299244 1 -0 1 -368 0 -5.8557806 1E-15 1.4415419267167138E-15 0 -369 0 -5.4592943 1E-15 1.4415419267167138E-15 0 -370 0 -3.8947306 1E-15 1.4415419267167138E-15 0 -371 0 -5.8557806 1E-15 1.4415419267167138E-15 0 -372 0 -4.2402444 1E-15 1.4415419267167138E-15 0 -373 0 -3.7544198 1E-15 1.4415419267167138E-15 0 -374 0 -4.71424 1E-15 1.4415419267167138E-15 0 -375 0 -6.442978 1E-15 1.4415419267167138E-15 0 -376 0 -5.95583 1E-15 1.4415419267167138E-15 0 -377 0 -6.0464916 1E-15 1.4415419267167138E-15 0 -378 0 -3.8038664 1E-15 1.4415419267167138E-15 0 -379 0 -2.2557268 1E-15 1.4415419267167138E-15 0 -380 0 -6.442978 1E-15 1.4415419267167138E-15 0 -381 1 10.076408 1 -0 1 -382 0 -3.5972104 1E-15 1.4415419267167138E-15 0 -383 0 -5.9689827 1E-15 1.4415419267167138E-15 0 -384 0 -5.9689827 1E-15 1.4415419267167138E-15 0 -385 0 -3.7061968 1E-15 1.4415419267167138E-15 0 -386 1 6.0875874 0.98 0.029146317580716615 1 -387 0 -2.33456 1E-15 1.4415419267167138E-15 0 -388 0 -5.2848067 1E-15 1.4415419267167138E-15 0 -389 0 -3.322411 1E-15 1.4415419267167138E-15 0 -390 0 -5.6504025 1E-15 1.4415419267167138E-15 0 -391 1 10.030338 1 -0 1 -392 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -393 0 -6.53364 1E-15 1.4415419267167138E-15 0 -394 0 -5.241206 1E-15 1.4415419267167138E-15 0 -395 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -396 0 -4.520691 1E-15 1.4415419267167138E-15 0 -397 0 -5.0209913 1E-15 1.4415419267167138E-15 0 -398 0 -4.683385 1E-15 1.4415419267167138E-15 0 -399 0 -5.7283545 1E-15 1.4415419267167138E-15 0 -400 1 10.056744 1 -0 1 -401 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -402 0 -3.2177973 1E-15 1.4415419267167138E-15 0 -403 0 -4.145746 1E-15 1.4415419267167138E-15 0 -404 0 -5.507669 1E-15 1.4415419267167138E-15 0 -405 0 -5.95583 1E-15 1.4415419267167138E-15 0 -406 0 -4.1128182 1E-15 1.4415419267167138E-15 0 -407 0 -5.95583 1E-15 1.4415419267167138E-15 0 -408 0 -3.6910605 1E-15 1.4415419267167138E-15 0 -409 0 -4.71424 1E-15 1.4415419267167138E-15 0 -410 0 -5.95583 1E-15 1.4415419267167138E-15 0 +322 0 -4.697876 1E-15 1.4415419267167138E-15 0 +323 1 5.534587 0.9814815 0.026967031375491075 1 +324 0 -5.690696 1E-15 1.4415419267167138E-15 0 +325 0 -4.4049735 1E-15 1.4415419267167138E-15 0 +326 1 4.4437723 0.9230769 0.11547721025399223 1 +327 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +328 1 3.1013708 0.85714287 0.22239240700456045 1 +329 1 7.778722 1 -0 1 +330 1 5.7779264 0.9814815 0.026967031375491075 1 +331 0 -3.389574 1E-15 1.4415419267167138E-15 0 +332 0 -3.2830477 1E-15 1.4415419267167138E-15 0 +333 1 4.790513 0.9814815 0.026967031375491075 1 +334 1 5.7627764 0.9814815 0.026967031375491075 1 +335 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +336 1 5.4697657 0.9814815 0.026967031375491075 1 +337 0 -5.690696 1E-15 1.4415419267167138E-15 0 +338 0 -6.255251 1E-15 1.4415419267167138E-15 0 +339 1 5.63153 0.9814815 0.026967031375491075 1 +340 1 6.586154 0.9814815 0.026967031375491075 1 +341 0 -5.690696 1E-15 1.4415419267167138E-15 0 +342 0 -6.1653986 1E-15 1.4415419267167138E-15 0 +343 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +344 1 10.009007 1 -0 1 +345 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +346 0 -3.4967813 1E-15 1.4415419267167138E-15 0 +347 0 -6.3551583 1E-15 1.4415419267167138E-15 0 +348 1 0.0005674362 0.6923077 0.53051467848041345 1 +349 1 3.947217 0.9230769 0.11547721025399223 1 +350 0 -4.0323086 1E-15 1.4415419267167138E-15 0 +351 0 -5.194286 1E-15 1.4415419267167138E-15 0 +352 0 0.16456318 0.6923077 1.7004398041324202 1 +353 1 8.388366 1 -0 1 +354 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +355 0 -4.4151993 1E-15 1.4415419267167138E-15 0 +356 1 -0.90889454 0.4 1.3219280733895313 0 +357 1 12.536469 1 -0 1 +358 1 5.258912 0.9814815 0.026967031375491075 1 +359 1 5.3155527 0.9814815 0.026967031375491075 1 +360 1 14.975728 1 -0 1 +361 1 6.1216927 0.9814815 0.026967031375491075 1 +362 0 -3.7176933 1E-15 1.4415419267167138E-15 0 +363 0 -2.2717314 0.083333336 0.12553088599255555 0 +364 0 -5.194286 1E-15 1.4415419267167138E-15 0 +365 0 -5.679842 1E-15 1.4415419267167138E-15 0 +366 1 13.38573 1 -0 1 +367 1 11.090333 1 -0 1 +368 0 -6.1007133 1E-15 1.4415419267167138E-15 0 +369 0 -5.6941557 1E-15 1.4415419267167138E-15 0 +370 0 -4.1029515 1E-15 1.4415419267167138E-15 0 +371 0 -6.1007133 1E-15 1.4415419267167138E-15 0 +372 0 -4.4064746 1E-15 1.4415419267167138E-15 0 +373 0 -3.868682 1E-15 1.4415419267167138E-15 0 +374 0 -4.9028845 1E-15 1.4415419267167138E-15 0 +375 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +376 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +377 0 -6.255251 1E-15 1.4415419267167138E-15 0 +378 0 -3.9458194 1E-15 1.4415419267167138E-15 0 +379 0 -2.4808912 1E-15 1.4415419267167138E-15 0 +380 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +381 1 10.022679 1 -0 1 +382 0 -3.7903118 1E-15 1.4415419267167138E-15 0 +383 0 -6.1653986 1E-15 1.4415419267167138E-15 0 +384 0 -6.1653986 1E-15 1.4415419267167138E-15 0 +385 0 -3.9392 1E-15 1.4415419267167138E-15 0 +386 1 5.9550896 0.9814815 0.026967031375491075 1 +387 0 -2.518508 1E-15 1.4415419267167138E-15 0 +388 0 -5.4870176 1E-15 1.4415419267167138E-15 0 +389 0 -3.4613237 1E-15 1.4415419267167138E-15 0 +390 0 -5.8543224 1E-15 1.4415419267167138E-15 0 +391 1 9.892915 1 -0 1 +392 0 -5.183432 1E-15 1.4415419267167138E-15 0 +393 0 -6.740807 1E-15 1.4415419267167138E-15 0 +394 0 -5.5028944 1E-15 1.4415419267167138E-15 0 +395 0 -5.183432 1E-15 1.4415419267167138E-15 0 +396 0 -4.697876 1E-15 1.4415419267167138E-15 0 +397 0 -5.172579 1E-15 1.4415419267167138E-15 0 +398 0 -4.8783455 1E-15 1.4415419267167138E-15 0 +399 0 -5.988451 1E-15 1.4415419267167138E-15 0 +400 1 9.690976 1 -0 1 +401 0 -5.668988 1E-15 1.4415419267167138E-15 0 +402 0 -3.4049926 1E-15 1.4415419267167138E-15 0 +403 0 -4.296591 1E-15 1.4415419267167138E-15 0 +404 0 -5.712862 1E-15 1.4415419267167138E-15 0 +405 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +406 0 -4.2942123 1E-15 1.4415419267167138E-15 0 +407 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +408 0 -3.9098177 1E-15 1.4415419267167138E-15 0 +409 0 -4.9028845 1E-15 1.4415419267167138E-15 0 +410 0 -6.1762524 1E-15 1.4415419267167138E-15 0 411 0 ? ? ? 0 -412 1 9.230707 1 -0 1 -413 0 -3.279101 1E-15 1.4415419267167138E-15 0 -414 1 6.7173805 0.98 0.029146317580716615 1 -415 0 -0.6668339 0.5 1 0 -416 1 8.809383 1 -0 1 -417 0 -5.95583 1E-15 1.4415419267167138E-15 0 -418 0 -1.8758616 0.071428575 0.10691520887754996 0 -419 0 -5.4421444 1E-15 1.4415419267167138E-15 0 -420 0 -2.5893164 1E-15 1.4415419267167138E-15 0 -421 1 11.824856 1 -0 1 -422 0 -2.8105893 1E-15 1.4415419267167138E-15 0 -423 0 -3.4726496 1E-15 1.4415419267167138E-15 0 -424 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -425 1 14.817663 1 -0 1 -426 0 -2.8241491 1E-15 1.4415419267167138E-15 0 -427 1 4.3530817 0.9285714 0.10691524360481655 1 -428 0 -5.95583 1E-15 1.4415419267167138E-15 0 -429 0 -5.4818344 1E-15 1.4415419267167138E-15 0 -430 0 -5.50395 1E-15 1.4415419267167138E-15 0 -431 0 -2.5834928 1E-15 1.4415419267167138E-15 0 -432 0 -3.8628197 1E-15 1.4415419267167138E-15 0 -433 0 -4.463106 1E-15 1.4415419267167138E-15 0 -434 0 5.008401 0.9285714 3.8073544061097437 1 -435 1 7.444332 1 -0 1 -436 1 3.841199 0.9285714 0.10691524360481655 1 -437 0 -5.0209913 1E-15 1.4415419267167138E-15 0 -438 0 -3.822938 1E-15 1.4415419267167138E-15 0 -439 0 -4.5469956 1E-15 1.4415419267167138E-15 0 -440 1 10.154686 1 -0 1 -441 0 -1.8604474 0.071428575 0.10691520887754996 0 -442 0 -4.9326286 1E-15 1.4415419267167138E-15 0 -443 0 -5.9932313 1E-15 1.4415419267167138E-15 0 -444 0 -2.442047 1E-15 1.4415419267167138E-15 0 -445 0 -5.9689827 1E-15 1.4415419267167138E-15 0 -446 0 -6.442978 1E-15 1.4415419267167138E-15 0 -447 0 -4.5469956 1E-15 1.4415419267167138E-15 0 -448 0 -6.53364 1E-15 1.4415419267167138E-15 0 -449 1 10.298004 1 -0 1 -450 0 -4.01365 1E-15 1.4415419267167138E-15 0 -451 0 -4.5469956 1E-15 1.4415419267167138E-15 0 -452 0 -4.8841314 1E-15 1.4415419267167138E-15 0 -453 1 8.777969 1 -0 1 -454 0 -5.6234684 1E-15 1.4415419267167138E-15 0 -455 1 0.81635284 0.6363636 0.65207672114864346 1 -456 1 10.487385 1 -0 1 -457 1 9.062628 1 -0 1 -458 0 -4.253397 1E-15 1.4415419267167138E-15 0 -459 0 -3.9597979 1E-15 1.4415419267167138E-15 0 -460 0 -3.93614 1E-15 1.4415419267167138E-15 0 -461 0 -3.6959996 1E-15 1.4415419267167138E-15 0 -462 0 -3.4621449 1E-15 1.4415419267167138E-15 0 -463 0 -4.823963 1E-15 1.4415419267167138E-15 0 -464 0 -5.0209913 1E-15 1.4415419267167138E-15 0 -465 1 9.7853985 1 -0 1 -466 1 9.541931 1 -0 1 -467 1 7.7145195 1 -0 1 -468 0 -5.0209913 1E-15 1.4415419267167138E-15 0 -469 0 -5.6622314 1E-15 1.4415419267167138E-15 0 -470 0 -5.0985007 1E-15 1.4415419267167138E-15 0 -471 0 -3.4621449 1E-15 1.4415419267167138E-15 0 -472 0 -4.163662 1E-15 1.4415419267167138E-15 0 -473 0 -5.0209913 1E-15 1.4415419267167138E-15 0 -474 0 -4.5469956 1E-15 1.4415419267167138E-15 0 -475 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -476 0 -4.7273927 1E-15 1.4415419267167138E-15 0 -477 0 -5.0209913 1E-15 1.4415419267167138E-15 0 -478 0 -4.4195695 1E-15 1.4415419267167138E-15 0 -479 1 8.32148 1 -0 1 -480 0 -4.6376576 1E-15 1.4415419267167138E-15 0 -481 0 -3.0822616 1E-15 1.4415419267167138E-15 0 -482 1 15.481422 1 -0 1 -483 1 10.906595 1 -0 1 -484 0 -4.253397 1E-15 1.4415419267167138E-15 0 -485 0 -5.0232906 1E-15 1.4415419267167138E-15 0 -486 0 -5.0985007 1E-15 1.4415419267167138E-15 0 -487 1 13.475906 1 -0 1 -488 1 1.3273249 0.8095238 0.30485456129516797 1 -489 1 -0.6232023 0.5 1 0 -490 0 -6.442978 1E-15 1.4415419267167138E-15 0 -491 1 6.6713343 0.98 0.029146317580716615 1 -492 0 -4.624505 1E-15 1.4415419267167138E-15 0 -493 1 9.948912 1 -0 1 -494 0 0.9629116 0.6363636 1.4594315756416352 1 -495 0 -5.0985007 1E-15 1.4415419267167138E-15 0 -496 0 -6.53364 1E-15 1.4415419267167138E-15 0 -497 0 -4.893565 1E-15 1.4415419267167138E-15 0 -498 0 -4.533843 1E-15 1.4415419267167138E-15 0 -499 0 -4.533843 1E-15 1.4415419267167138E-15 0 -500 0 -3.0987039 1E-15 1.4415419267167138E-15 0 -501 0 -4.533843 1E-15 1.4415419267167138E-15 0 -502 0 -4.2284155 1E-15 1.4415419267167138E-15 0 -503 0 -4.046695 1E-15 1.4415419267167138E-15 0 -504 0 -6.442978 1E-15 1.4415419267167138E-15 0 -505 0 -5.2401342 1E-15 1.4415419267167138E-15 0 -506 1 10.447666 1 -0 1 -507 0 -5.0937376 1E-15 1.4415419267167138E-15 0 -508 0 -4.5469956 1E-15 1.4415419267167138E-15 0 -509 0 -5.9689827 1E-15 1.4415419267167138E-15 0 -510 0 -6.442978 1E-15 1.4415419267167138E-15 0 -511 0 -4.0598474 1E-15 1.4415419267167138E-15 0 -512 0 -4.5469956 1E-15 1.4415419267167138E-15 0 -513 0 -5.0985007 1E-15 1.4415419267167138E-15 0 -514 1 10.719856 1 -0 1 -515 1 8.648027 1 -0 1 -516 0 -6.53364 1E-15 1.4415419267167138E-15 0 -517 0 -6.0464916 1E-15 1.4415419267167138E-15 0 -518 0 -4.8959603 1E-15 1.4415419267167138E-15 0 -519 1 6.535844 0.98 0.029146317580716615 1 -520 0 -6.3523164 1E-15 1.4415419267167138E-15 0 -521 0 -4.9303293 1E-15 1.4415419267167138E-15 0 -522 1 5.502533 0.98 0.029146317580716615 1 -523 1 7.699238 1 -0 1 -524 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -525 0 -5.189559 1E-15 1.4415419267167138E-15 0 -526 0 -5.0209913 1E-15 1.4415419267167138E-15 0 -527 0 -4.046695 1E-15 1.4415419267167138E-15 0 -528 0 -3.1803741 1E-15 1.4415419267167138E-15 0 -529 0 -4.624505 1E-15 1.4415419267167138E-15 0 -530 1 6.515936 0.98 0.029146317580716615 1 -531 0 -4.1128182 1E-15 1.4415419267167138E-15 0 -532 0 -5.559344 1E-15 1.4415419267167138E-15 0 -533 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -534 0 -5.4818344 1E-15 1.4415419267167138E-15 0 -535 0 -4.5760565 1E-15 1.4415419267167138E-15 0 -536 0 -3.5726995 1E-15 1.4415419267167138E-15 0 -537 0 -3.279101 1E-15 1.4415419267167138E-15 0 -538 0 -4.533843 1E-15 1.4415419267167138E-15 0 -539 0 -3.5858517 1E-15 1.4415419267167138E-15 0 -540 0 -3.6101003 1E-15 1.4415419267167138E-15 0 -541 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -542 0 -4.2921433 1E-15 1.4415419267167138E-15 0 -543 0 -4.533843 1E-15 1.4415419267167138E-15 0 -544 0 -4.589209 1E-15 1.4415419267167138E-15 0 -545 0 -4.0598474 1E-15 1.4415419267167138E-15 0 -546 1 11.390259 1 -0 1 -547 0 -6.0596447 1E-15 1.4415419267167138E-15 0 -548 0 -5.5856485 1E-15 1.4415419267167138E-15 0 -549 1 6.3187485 0.98 0.029146317580716615 1 -550 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -551 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -552 0 -3.1100621 1E-15 1.4415419267167138E-15 0 -553 0 -1.7353668 0.071428575 0.10691520887754996 0 -554 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -555 0 -1.9254999 0.071428575 0.10691520887754996 0 -556 0 -3.4240565 1E-15 1.4415419267167138E-15 0 -557 0 -3.93614 1E-15 1.4415419267167138E-15 0 -558 0 -5.4818344 1E-15 1.4415419267167138E-15 0 -559 0 -4.0598474 1E-15 1.4415419267167138E-15 0 -560 0 -3.5726995 1E-15 1.4415419267167138E-15 0 -561 0 -3.5726995 1E-15 1.4415419267167138E-15 0 -562 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -563 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -564 0 -3.8496675 1E-15 1.4415419267167138E-15 0 -565 1 12.086092 1 -0 1 -566 0 -4.2270923 1E-15 1.4415419267167138E-15 0 -567 0 -3.6343493 1E-15 1.4415419267167138E-15 0 -568 1 4.1473055 0.9285714 0.10691524360481655 1 -569 1 10.713882 1 -0 1 -570 1 8.017664 1 -0 1 -571 1 11.034658 1 -0 1 -572 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -573 0 -5.95583 1E-15 1.4415419267167138E-15 0 -574 1 5.950967 0.98 0.029146317580716615 1 -575 0 -3.279101 1E-15 1.4415419267167138E-15 0 -576 0 -4.0598474 1E-15 1.4415419267167138E-15 0 -577 0 -5.95583 1E-15 1.4415419267167138E-15 0 -578 0 -5.95583 1E-15 1.4415419267167138E-15 0 -579 0 -5.4686823 1E-15 1.4415419267167138E-15 0 -580 0 -3.7662487 1E-15 1.4415419267167138E-15 0 -581 1 8.417797 1 -0 1 -582 1 7.90798 1 -0 1 -583 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -584 0 -2.9291954 1E-15 1.4415419267167138E-15 0 -585 0 -6.442978 1E-15 1.4415419267167138E-15 0 -586 1 13.98102 1 -0 1 -587 0 -3.8628197 1E-15 1.4415419267167138E-15 0 -588 1 5.463169 0.98 0.029146317580716615 1 -589 0 -4.5469956 1E-15 1.4415419267167138E-15 0 -590 1 3.9684038 0.9285714 0.10691524360481655 1 -591 1 5.966527 0.98 0.029146317580716615 1 -592 1 5.7801704 0.98 0.029146317580716615 1 -593 0 -4.253397 1E-15 1.4415419267167138E-15 0 -594 1 5.101776 0.98 0.029146317580716615 1 -595 0 -4.0598474 1E-15 1.4415419267167138E-15 0 -596 0 -4.2402444 1E-15 1.4415419267167138E-15 0 -597 0 -2.9855018 1E-15 1.4415419267167138E-15 0 -598 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -599 0 -3.6294346 1E-15 1.4415419267167138E-15 0 -600 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -601 0 -6.0464916 1E-15 1.4415419267167138E-15 0 -602 0 -4.533843 1E-15 1.4415419267167138E-15 0 -603 1 4.8058825 0.9285714 0.10691524360481655 1 -604 1 6.1928034 0.98 0.029146317580716615 1 -605 1 9.95545 1 -0 1 -606 0 -4.715564 1E-15 1.4415419267167138E-15 0 -607 0 -6.442978 1E-15 1.4415419267167138E-15 0 -608 1 11.148426 1 -0 1 -609 0 -4.5469956 1E-15 1.4415419267167138E-15 0 -610 1 8.926191 1 -0 1 -611 1 6.9109592 1 -0 1 -612 1 16.893515 1 -0 1 -613 0 -5.226982 1E-15 1.4415419267167138E-15 0 -614 0 -5.5724964 1E-15 1.4415419267167138E-15 0 -615 0 -3.9466453 1E-15 1.4415419267167138E-15 0 -616 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +412 1 8.967401 1 -0 1 +413 0 -3.424508 1E-15 1.4415419267167138E-15 0 +414 1 6.574872 0.9814815 0.026967031375491075 1 +415 0 -0.8179264 0.4 0.73696560849809378 0 +416 1 8.593532 1 -0 1 +417 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +418 0 -2.0245438 0.083333336 0.12553088599255555 0 +419 0 -5.7038784 1E-15 1.4415419267167138E-15 0 +420 0 -2.7462745 1E-15 1.4415419267167138E-15 0 +421 1 11.613648 1 -0 1 +422 0 -3.0301971 1E-15 1.4415419267167138E-15 0 +423 0 -3.6295166 1E-15 1.4415419267167138E-15 0 +424 0 -5.668988 1E-15 1.4415419267167138E-15 0 +425 1 14.596736 1 -0 1 +426 0 -3.0489054 1E-15 1.4415419267167138E-15 0 +427 1 4.065505 0.9230769 0.11547721025399223 1 +428 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +429 0 -5.679842 1E-15 1.4415419267167138E-15 0 +430 0 -5.7589035 1E-15 1.4415419267167138E-15 0 +431 0 -2.7103744 1E-15 1.4415419267167138E-15 0 +432 0 -3.9977875 1E-15 1.4415419267167138E-15 0 +433 0 -4.6323185 1E-15 1.4415419267167138E-15 0 +434 0 4.7780943 0.9230769 3.7004398041324205 1 +435 1 7.255061 1 -0 1 +436 1 3.4924908 0.9230769 0.11547721025399223 1 +437 0 -5.172579 1E-15 1.4415419267167138E-15 0 +438 0 -3.9676237 1E-15 1.4415419267167138E-15 0 +439 0 -4.6761684 1E-15 1.4415419267167138E-15 0 +440 1 10.246327 1 -0 1 +441 0 -1.9451714 0.083333336 0.12553088599255555 0 +442 0 -5.1030173 1E-15 1.4415419267167138E-15 0 +443 0 -6.204343 1E-15 1.4415419267167138E-15 0 +444 0 -2.501964 1E-15 1.4415419267167138E-15 0 +445 0 -6.1653986 1E-15 1.4415419267167138E-15 0 +446 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +447 0 -4.6761684 1E-15 1.4415419267167138E-15 0 +448 0 -6.740807 1E-15 1.4415419267167138E-15 0 +449 1 10.062795 1 -0 1 +450 0 -4.122161 1E-15 1.4415419267167138E-15 0 +451 0 -4.6761684 1E-15 1.4415419267167138E-15 0 +452 0 -5.0251284 1E-15 1.4415419267167138E-15 0 +453 1 8.612309 1 -0 1 +454 0 -5.8618755 1E-15 1.4415419267167138E-15 0 +455 1 0.8015022 0.6923077 0.53051467848041345 1 +456 1 10.460095 1 -0 1 +457 1 8.789279 1 -0 1 +458 0 -4.395621 1E-15 1.4415419267167138E-15 0 +459 0 -4.115073 1E-15 1.4415419267167138E-15 0 +460 0 -4.0323086 1E-15 1.4415419267167138E-15 0 +461 0 -3.8391252 1E-15 1.4415419267167138E-15 0 +462 0 -3.5358982 1E-15 1.4415419267167138E-15 0 +463 0 -4.9797544 1E-15 1.4415419267167138E-15 0 +464 0 -5.172579 1E-15 1.4415419267167138E-15 0 +465 1 9.567504 1 -0 1 +466 1 9.408446 1 -0 1 +467 1 7.540141 1 -0 1 +468 0 -5.172579 1E-15 1.4415419267167138E-15 0 +469 0 -5.8957047 1E-15 1.4415419267167138E-15 0 +470 0 -5.2624307 1E-15 1.4415419267167138E-15 0 +471 0 -3.5358982 1E-15 1.4415419267167138E-15 0 +472 0 -4.2587566 1E-15 1.4415419267167138E-15 0 +473 0 -5.172579 1E-15 1.4415419267167138E-15 0 +474 0 -4.6761684 1E-15 1.4415419267167138E-15 0 +475 0 -5.668988 1E-15 1.4415419267167138E-15 0 +476 0 -4.8920307 1E-15 1.4415419267167138E-15 0 +477 0 -5.172579 1E-15 1.4415419267167138E-15 0 +478 0 -4.563906 1E-15 1.4415419267167138E-15 0 +479 1 8.153737 1 -0 1 +480 0 -4.755167 1E-15 1.4415419267167138E-15 0 +481 0 -3.205307 1E-15 1.4415419267167138E-15 0 +482 1 15.148742 1 -0 1 +483 1 10.712245 1 -0 1 +484 0 -4.395621 1E-15 1.4415419267167138E-15 0 +485 0 -5.182016 1E-15 1.4415419267167138E-15 0 +486 0 -5.2624307 1E-15 1.4415419267167138E-15 0 +487 1 13.332547 1 -0 1 +488 1 1.1748514 0.8 0.32192807338953117 1 +489 1 -0.7666216 0.6923077 0.53051467848041345 0 +490 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +491 1 6.429366 0.9814815 0.026967031375491075 1 +492 0 -4.766021 1E-15 1.4415419267167138E-15 0 +493 1 9.704236 1 -0 1 +494 0 0.97516537 0.6923077 1.7004398041324202 1 +495 0 -5.2624307 1E-15 1.4415419267167138E-15 0 +496 0 -6.740807 1E-15 1.4415419267167138E-15 0 +497 0 -5.060316 1E-15 1.4415419267167138E-15 0 +498 0 -4.687022 1E-15 1.4415419267167138E-15 0 +499 0 -4.687022 1E-15 1.4415419267167138E-15 0 +500 0 -3.2086458 1E-15 1.4415419267167138E-15 0 +501 0 -4.687022 1E-15 1.4415419267167138E-15 0 +502 0 -4.3650923 1E-15 1.4415419267167138E-15 0 +503 0 -4.2014656 1E-15 1.4415419267167138E-15 0 +504 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +505 0 -5.4444637 1E-15 1.4415419267167138E-15 0 +506 1 10.169904 1 -0 1 +507 0 -5.289411 1E-15 1.4415419267167138E-15 0 +508 0 -4.6761684 1E-15 1.4415419267167138E-15 0 +509 0 -6.1653986 1E-15 1.4415419267167138E-15 0 +510 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +511 0 -4.190612 1E-15 1.4415419267167138E-15 0 +512 0 -4.6761684 1E-15 1.4415419267167138E-15 0 +513 0 -5.2624307 1E-15 1.4415419267167138E-15 0 +514 1 10.608193 1 -0 1 +515 1 8.427064 1 -0 1 +516 0 -6.740807 1E-15 1.4415419267167138E-15 0 +517 0 -6.255251 1E-15 1.4415419267167138E-15 0 +518 0 -5.066511 1E-15 1.4415419267167138E-15 0 +519 1 6.2683277 0.9814815 0.026967031375491075 1 +520 0 -6.58281 1E-15 1.4415419267167138E-15 0 +521 0 -5.09358 1E-15 1.4415419267167138E-15 0 +522 1 5.4384966 0.9814815 0.026967031375491075 1 +523 1 7.66887 1 -0 1 +524 0 -5.183432 1E-15 1.4415419267167138E-15 0 +525 0 -5.3470583 1E-15 1.4415419267167138E-15 0 +526 0 -5.172579 1E-15 1.4415419267167138E-15 0 +527 0 -4.2014656 1E-15 1.4415419267167138E-15 0 +528 0 -3.296733 1E-15 1.4415419267167138E-15 0 +529 0 -4.766021 1E-15 1.4415419267167138E-15 0 +530 1 6.384506 0.9814815 0.026967031375491075 1 +531 0 -4.2942123 1E-15 1.4415419267167138E-15 0 +532 0 -5.7696943 1E-15 1.4415419267167138E-15 0 +533 0 -5.183432 1E-15 1.4415419267167138E-15 0 +534 0 -5.679842 1E-15 1.4415419267167138E-15 0 +535 0 -4.8076706 1E-15 1.4415419267167138E-15 0 +536 0 -3.7050557 1E-15 1.4415419267167138E-15 0 +537 0 -3.424508 1E-15 1.4415419267167138E-15 0 +538 0 -4.687022 1E-15 1.4415419267167138E-15 0 +539 0 -3.694202 1E-15 1.4415419267167138E-15 0 +540 0 -3.7331467 1E-15 1.4415419267167138E-15 0 +541 0 -5.668988 1E-15 1.4415419267167138E-15 0 +542 0 -4.451644 1E-15 1.4415419267167138E-15 0 +543 0 -4.687022 1E-15 1.4415419267167138E-15 0 +544 0 -4.796817 1E-15 1.4415419267167138E-15 0 +545 0 -4.190612 1E-15 1.4415419267167138E-15 0 +546 1 11.1768465 1 -0 1 +547 0 -6.244397 1E-15 1.4415419267167138E-15 0 +548 0 -5.747987 1E-15 1.4415419267167138E-15 0 +549 1 6.1616106 0.9814815 0.026967031375491075 1 +550 0 -5.183432 1E-15 1.4415419267167138E-15 0 +551 0 -5.690696 1E-15 1.4415419267167138E-15 0 +552 0 -3.3047552 1E-15 1.4415419267167138E-15 0 +553 0 -1.9243431 0.083333336 0.12553088599255555 0 +554 0 -5.668988 1E-15 1.4415419267167138E-15 0 +555 0 -2.141615 0.083333336 0.12553088599255555 0 +556 0 -3.5548706 1E-15 1.4415419267167138E-15 0 +557 0 -4.0323086 1E-15 1.4415419267167138E-15 0 +558 0 -5.679842 1E-15 1.4415419267167138E-15 0 +559 0 -4.190612 1E-15 1.4415419267167138E-15 0 +560 0 -3.7050557 1E-15 1.4415419267167138E-15 0 +561 0 -3.7050557 1E-15 1.4415419267167138E-15 0 +562 0 -5.690696 1E-15 1.4415419267167138E-15 0 +563 0 -5.183432 1E-15 1.4415419267167138E-15 0 +564 0 -4.0086412 1E-15 1.4415419267167138E-15 0 +565 1 11.879315 1 -0 1 +566 0 -4.417328 1E-15 1.4415419267167138E-15 0 +567 0 -3.772091 1E-15 1.4415419267167138E-15 0 +568 1 3.9970121 0.9230769 0.11547721025399223 1 +569 1 10.578958 1 -0 1 +570 1 7.813756 1 -0 1 +571 1 10.822152 1 -0 1 +572 0 -5.183432 1E-15 1.4415419267167138E-15 0 +573 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +574 1 5.7168713 0.9814815 0.026967031375491075 1 +575 0 -3.424508 1E-15 1.4415419267167138E-15 0 +576 0 -4.190612 1E-15 1.4415419267167138E-15 0 +577 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +578 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +579 0 -5.690696 1E-15 1.4415419267167138E-15 0 +580 0 -3.9100647 1E-15 1.4415419267167138E-15 0 +581 1 8.170806 1 -0 1 +582 1 7.6789007 1 -0 1 +583 0 -5.668988 1E-15 1.4415419267167138E-15 0 +584 0 -2.9875202 1E-15 1.4415419267167138E-15 0 +585 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +586 1 13.784331 1 -0 1 +587 0 -3.9977875 1E-15 1.4415419267167138E-15 0 +588 1 5.322174 0.9814815 0.026967031375491075 1 +589 0 -4.6761684 1E-15 1.4415419267167138E-15 0 +590 1 3.6173382 0.9230769 0.11547721025399223 1 +591 1 5.7292347 0.9814815 0.026967031375491075 1 +592 1 5.8126287 0.9814815 0.026967031375491075 1 +593 0 -4.395621 1E-15 1.4415419267167138E-15 0 +594 1 4.848382 0.9814815 0.026967031375491075 1 +595 0 -4.190612 1E-15 1.4415419267167138E-15 0 +596 0 -4.4064746 1E-15 1.4415419267167138E-15 0 +597 0 -3.14396 1E-15 1.4415419267167138E-15 0 +598 0 -5.183432 1E-15 1.4415419267167138E-15 0 +599 0 -3.801262 1E-15 1.4415419267167138E-15 0 +600 0 -5.183432 1E-15 1.4415419267167138E-15 0 +601 0 -6.255251 1E-15 1.4415419267167138E-15 0 +602 0 -4.687022 1E-15 1.4415419267167138E-15 0 +603 1 4.6091757 0.9230769 0.11547721025399223 1 +604 1 6.150976 0.9814815 0.026967031375491075 1 +605 1 9.757047 1 -0 1 +606 0 -4.850649 1E-15 1.4415419267167138E-15 0 +607 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +608 1 10.875676 1 -0 1 +609 0 -4.6761684 1E-15 1.4415419267167138E-15 0 +610 1 9.026143 1 -0 1 +611 1 6.6211424 0.9814815 0.026967031375491075 1 +612 1 16.623665 1 -0 1 +613 0 -5.4553175 1E-15 1.4415419267167138E-15 0 +614 0 -5.7588406 1E-15 1.4415419267167138E-15 0 +615 0 -4.125927 1E-15 1.4415419267167138E-15 0 +616 0 -5.183432 1E-15 1.4415419267167138E-15 0 617 0 ? ? ? 0 -618 0 -4.533843 1E-15 1.4415419267167138E-15 0 -619 0 -4.0598474 1E-15 1.4415419267167138E-15 0 -620 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -621 0 0.3560543 0.6363636 1.4594315756416352 1 -622 0 -2.2074018 0.071428575 0.10691520887754996 0 -623 0 -6.442978 1E-15 1.4415419267167138E-15 0 -624 0 -3.8450818 1E-15 1.4415419267167138E-15 0 -625 0 -3.4678864 1E-15 1.4415419267167138E-15 0 -626 1 5.760166 0.98 0.029146317580716615 1 -627 0 -4.1699953 1E-15 1.4415419267167138E-15 0 -628 0 -5.9689827 1E-15 1.4415419267167138E-15 0 -629 0 -5.0209913 1E-15 1.4415419267167138E-15 0 -630 0 -3.358376 1E-15 1.4415419267167138E-15 0 -631 0 -4.0598474 1E-15 1.4415419267167138E-15 0 -632 0 -6.442978 1E-15 1.4415419267167138E-15 0 -633 1 4.3299494 0.9285714 0.10691524360481655 1 -634 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -635 0 -4.6141906 1E-15 1.4415419267167138E-15 0 -636 1 10.12752 1 -0 1 -637 0 -2.4650183 1E-15 1.4415419267167138E-15 0 -638 0 -5.0209913 1E-15 1.4415419267167138E-15 0 -639 0 -3.93614 1E-15 1.4415419267167138E-15 0 -640 0 -4.4101357 1E-15 1.4415419267167138E-15 0 -641 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -642 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -643 0 -6.442978 1E-15 1.4415419267167138E-15 0 -644 0 -5.9689827 1E-15 1.4415419267167138E-15 0 -645 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -646 0 -6.021953 1E-15 1.4415419267167138E-15 0 -647 0 -5.832123 1E-15 1.4415419267167138E-15 0 -648 1 12.2362 1 -0 1 -649 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -650 0 -3.7496572 1E-15 1.4415419267167138E-15 0 -651 0 -5.2175484 1E-15 1.4415419267167138E-15 0 -652 0 -3.8628197 1E-15 1.4415419267167138E-15 0 -653 0 -4.533843 1E-15 1.4415419267167138E-15 0 -654 0 -4.520691 1E-15 1.4415419267167138E-15 0 -655 0 -5.0078387 1E-15 1.4415419267167138E-15 0 -656 0 -4.0598474 1E-15 1.4415419267167138E-15 0 -657 0 -0.4869156 0.5 1 0 -658 1 9.086258 1 -0 1 -659 0 -6.442978 1E-15 1.4415419267167138E-15 0 -660 0 -5.95583 1E-15 1.4415419267167138E-15 0 -661 0 -4.046695 1E-15 1.4415419267167138E-15 0 -662 0 -5.3686323 1E-15 1.4415419267167138E-15 0 -663 0 -5.3686323 1E-15 1.4415419267167138E-15 0 -664 0 -4.3969836 1E-15 1.4415419267167138E-15 0 -665 0 -6.442978 1E-15 1.4415419267167138E-15 0 -666 0 -3.496932 1E-15 1.4415419267167138E-15 0 -667 0 -4.520691 1E-15 1.4415419267167138E-15 0 -668 1 2.804366 0.8095238 0.30485456129516797 1 -669 1 8.147337 1 -0 1 -670 1 6.4856215 0.98 0.029146317580716615 1 -671 0 -4.087837 1E-15 1.4415419267167138E-15 0 -672 0 -4.9946866 1E-15 1.4415419267167138E-15 0 -673 0 -3.9078827 1E-15 1.4415419267167138E-15 0 -674 0 -5.95583 1E-15 1.4415419267167138E-15 0 -675 0 -4.140195 1E-15 1.4415419267167138E-15 0 -676 0 -5.6622314 1E-15 1.4415419267167138E-15 0 -677 0 -4.5469956 1E-15 1.4415419267167138E-15 0 -678 0 -6.442978 1E-15 1.4415419267167138E-15 0 -679 0 -5.9689827 1E-15 1.4415419267167138E-15 0 -680 1 16.78001 1 -0 1 -681 1 9.801079 1 -0 1 -682 0 -3.3756719 1E-15 1.4415419267167138E-15 0 -683 0 -6.442978 1E-15 1.4415419267167138E-15 0 -684 0 -6.442978 1E-15 1.4415419267167138E-15 0 -685 0 -6.442978 1E-15 1.4415419267167138E-15 0 -686 0 -6.442978 1E-15 1.4415419267167138E-15 0 -687 0 -4.613783 1E-15 1.4415419267167138E-15 0 -688 0 -5.0209913 1E-15 1.4415419267167138E-15 0 -689 0 -4.177704 1E-15 1.4415419267167138E-15 0 -690 0 -5.832123 1E-15 1.4415419267167138E-15 0 -691 1 4.4967804 0.9285714 0.10691524360481655 1 -692 0 -5.4949865 1E-15 1.4415419267167138E-15 0 -693 0 -4.684228 1E-15 1.4415419267167138E-15 0 -694 0 -4.9545784 1E-15 1.4415419267167138E-15 0 -695 0 -5.9689827 1E-15 1.4415419267167138E-15 0 -696 1 6.7127876 0.98 0.029146317580716615 1 -697 1 5.064643 0.98 0.029146317580716615 1 -698 1 6.1036224 0.98 0.029146317580716615 1 +618 0 -4.687022 1E-15 1.4415419267167138E-15 0 +619 0 -4.190612 1E-15 1.4415419267167138E-15 0 +620 0 -5.183432 1E-15 1.4415419267167138E-15 0 +621 0 0.21396732 0.6923077 1.7004398041324202 1 +622 0 -2.2733846 0.083333336 0.12553088599255555 0 +623 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +624 0 -3.9476805 1E-15 1.4415419267167138E-15 0 +625 0 -3.656497 1E-15 1.4415419267167138E-15 0 +626 1 5.46276 0.9814815 0.026967031375491075 1 +627 0 -4.335535 1E-15 1.4415419267167138E-15 0 +628 0 -6.1653986 1E-15 1.4415419267167138E-15 0 +629 0 -5.172579 1E-15 1.4415419267167138E-15 0 +630 0 -3.506401 1E-15 1.4415419267167138E-15 0 +631 0 -4.190612 1E-15 1.4415419267167138E-15 0 +632 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +633 1 4.123639 0.9230769 0.11547721025399223 1 +634 0 -5.668988 1E-15 1.4415419267167138E-15 0 +635 0 -4.827346 1E-15 1.4415419267167138E-15 0 +636 1 9.884419 1 -0 1 +637 0 -2.6317916 1E-15 1.4415419267167138E-15 0 +638 0 -5.172579 1E-15 1.4415419267167138E-15 0 +639 0 -4.0323086 1E-15 1.4415419267167138E-15 0 +640 0 -4.528719 1E-15 1.4415419267167138E-15 0 +641 0 -5.183432 1E-15 1.4415419267167138E-15 0 +642 0 -5.183432 1E-15 1.4415419267167138E-15 0 +643 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +644 0 -6.1653986 1E-15 1.4415419267167138E-15 0 +645 0 -5.183432 1E-15 1.4415419267167138E-15 0 +646 0 -6.268999 1E-15 1.4415419267167138E-15 0 +647 0 -6.0179486 1E-15 1.4415419267167138E-15 0 +648 1 11.790706 1 -0 1 +649 0 -5.183432 1E-15 1.4415419267167138E-15 0 +650 0 -3.8956623 1E-15 1.4415419267167138E-15 0 +651 0 -5.4201303 1E-15 1.4415419267167138E-15 0 +652 0 -3.9977875 1E-15 1.4415419267167138E-15 0 +653 0 -4.687022 1E-15 1.4415419267167138E-15 0 +654 0 -4.697876 1E-15 1.4415419267167138E-15 0 +655 0 -5.183432 1E-15 1.4415419267167138E-15 0 +656 0 -4.190612 1E-15 1.4415419267167138E-15 0 +657 0 -0.8079653 0.4 0.73696560849809378 0 +658 1 8.931214 1 -0 1 +659 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +660 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +661 0 -4.2014656 1E-15 1.4415419267167138E-15 0 +662 0 -5.615157 1E-15 1.4415419267167138E-15 0 +663 0 -5.615157 1E-15 1.4415419267167138E-15 0 +664 0 -4.5395722 1E-15 1.4415419267167138E-15 0 +665 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +666 0 -3.645947 1E-15 1.4415419267167138E-15 0 +667 0 -4.697876 1E-15 1.4415419267167138E-15 0 +668 1 2.434762 0.8 0.32192807338953117 1 +669 1 7.828244 1 -0 1 +670 1 6.2092333 0.9814815 0.026967031375491075 1 +671 0 -4.2636833 1E-15 1.4415419267167138E-15 0 +672 0 -5.194286 1E-15 1.4415419267167138E-15 0 +673 0 -4.0920978 1E-15 1.4415419267167138E-15 0 +674 0 -6.1762524 1E-15 1.4415419267167138E-15 0 +675 0 -4.3309355 1E-15 1.4415419267167138E-15 0 +676 0 -5.8957047 1E-15 1.4415419267167138E-15 0 +677 0 -4.6761684 1E-15 1.4415419267167138E-15 0 +678 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +679 0 -6.1653986 1E-15 1.4415419267167138E-15 0 +680 1 16.591042 1 -0 1 +681 1 9.56899 1 -0 1 +682 0 -3.5122313 1E-15 1.4415419267167138E-15 0 +683 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +684 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +685 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +686 0 -6.6618085 1E-15 1.4415419267167138E-15 0 +687 0 -4.7977834 1E-15 1.4415419267167138E-15 0 +688 0 -5.172579 1E-15 1.4415419267167138E-15 0 +689 0 -4.4534364 1E-15 1.4415419267167138E-15 0 +690 0 -6.0179486 1E-15 1.4415419267167138E-15 0 +691 1 4.169842 0.9230769 0.11547721025399223 1 +692 0 -5.668988 1E-15 1.4415419267167138E-15 0 +693 0 -4.8679504 1E-15 1.4415419267167138E-15 0 +694 0 -5.132524 1E-15 1.4415419267167138E-15 0 +695 0 -6.1653986 1E-15 1.4415419267167138E-15 0 +696 1 6.262699 0.9814815 0.026967031375491075 1 +697 1 4.767309 0.9230769 0.11547721025399223 1 +698 1 5.8011494 0.9814815 0.026967031375491075 1 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-out.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-out.txt index d12e4da2b0..d3a552f3aa 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-out.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-out.txt @@ -8,28 +8,28 @@ Confusion table ||====================== PREDICTED || positive | negative | Recall TRUTH ||====================== - positive || 234 | 5 | 0.9791 + positive || 233 | 6 | 0.9749 negative || 12 | 432 | 0.9730 ||====================== -Precision || 0.9512 | 0.9886 | -OVERALL 0/1 ACCURACY: 0.975110 -LOG LOSS/instance: 0.120617 +Precision || 0.9510 | 0.9863 | +OVERALL 0/1 ACCURACY: 0.973646 +LOG LOSS/instance: 0.120727 Test-set entropy (prior Log-Loss/instance): 0.934003 -LOG-LOSS REDUCTION (RIG): 0.870860 +LOG-LOSS REDUCTION (RIG): 0.870742 AUC: 0.996146 OVERALL RESULTS --------------------------------------- AUC: 0.996146 (0.0000) -Accuracy: 0.975110 (0.0000) -Positive precision: 0.951220 (0.0000) -Positive recall: 0.979079 (0.0000) -Negative precision: 0.988558 (0.0000) +Accuracy: 0.973646 (0.0000) +Positive precision: 0.951020 (0.0000) +Positive recall: 0.974895 (0.0000) +Negative precision: 0.986301 (0.0000) Negative recall: 0.972973 (0.0000) -Log-loss: 0.120617 (0.0000) -Log-loss reduction: 0.870860 (0.0000) -F1 Score: 0.964948 (0.0000) -AUPRC: 0.992065 (0.0000) +Log-loss: 0.120727 (0.0000) +Log-loss reduction: 0.870742 (0.0000) +F1 Score: 0.962810 (0.0000) +AUPRC: 0.992010 (0.0000) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-rp.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-rp.txt index 0860eb7030..b59497818f 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-rp.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-rp.txt @@ -1,4 +1,4 @@ AveragedPerceptron AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings -0.996146 0.97511 0.95122 0.979079 0.988558 0.972973 0.120617 0.87086 0.964948 0.992065 AveragedPerceptron %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=AveragedPerceptron numcali=200 dout=%Output% data=%Data% out=%Output% seed=1 +0.996146 0.973646 0.95102 0.974895 0.986301 0.972973 0.120727 0.870742 0.96281 0.99201 AveragedPerceptron %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=AveragedPerceptron numcali=200 dout=%Output% data=%Data% out=%Output% seed=1 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom.txt index 55a840eda7..bd34063997 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom.txt @@ -1,700 +1,700 @@ Instance Label Score Probability Log-loss Assigned -0 0 -3.5726995 0.047724236 0.070548678886274141 0 -1 0 3.6101465 0.9209522 3.6611308948912158 1 -2 0 -4.070944 0.033202175 0.04871386747975371 0 -3 0 2.470542 0.83073896 2.5626781731328263 1 -4 0 -3.4358397 0.052670274 0.078061439656217479 0 -5 1 12.382595 0.99988943 0.0001595227277818751 1 -6 0 -1.4209604 0.20405398 0.32925750806385767 0 -7 0 -4.701088 0.020848805 0.03039644515807061 0 -8 0 -4.6745405 0.021263905 0.031008188656201943 0 -9 0 -4.406417 0.025935685 0.037911062322905087 0 -10 0 -5.559344 0.010982096 0.015931457121809788 0 -11 0 -5.4818344 0.011639432 0.016890641442649377 0 -12 1 -0.14206886 0.40347567 1.3094464075888503 0 -13 0 -4.5691886 0.022992346 0.033558231039772665 0 -14 1 9.321611 0.998874 0.0016253773809814112 1 -15 1 1.3856993 0.6830734 0.54988747944718552 1 -16 0 -4.533843 0.02360242 0.034459376167277234 0 -17 0 -4.046695 0.033797733 0.049602857843819056 0 -18 1 7.8903713 0.9966727 0.0048082975514731332 1 -19 0 -3.0987039 0.06699134 0.10003761775121778 0 -20 1 7.528511 0.9956263 0.0063237968594455725 1 -21 1 7.875206 0.9966343 0.0048638619113426269 1 -22 0 -5.0078387 0.016592305 0.02413845106369562 0 +0 0 -3.7050557 0.049417876 0.073116822887907204 0 +1 0 3.4661026 0.92318994 3.7025608769012859 1 +2 0 -4.24041 0.03346807 0.04911070335781155 0 +3 0 2.1493654 0.8156323 2.4393420399528236 1 +4 0 -3.5576057 0.054948628 0.081535339956679373 0 +5 1 12.155442 0.99988633 0.0001639947780942108 1 +6 0 -1.573504 0.20770586 0.33589195471326805 0 +7 0 -4.9137383 0.020348059 0.029658827501287101 0 +8 0 -4.903472 0.02050398 0.029888463292459169 0 +9 0 -4.57476 0.02616267 0.038247291919649574 0 +10 0 -5.7696943 0.010729991 0.01556375426175181 0 +11 0 -5.679842 0.011478641 0.016655956359413658 0 +12 1 -0.2966156 0.4086404 1.2910962014345189 0 +13 0 -4.7757645 0.02254403 0.032896378081435619 0 +14 1 9.220881 0.9989466 0.0015205256904994135 1 +15 1 1.1802778 0.6794937 0.55746786622082922 1 +16 0 -4.687022 0.02407708 0.035160889311831381 0 +17 0 -4.2014656 0.034437597 0.050558592955505986 0 +18 1 7.81973 0.99695474 0.0044000859180100124 1 +19 0 -3.2086458 0.07043917 0.10537881324670834 0 +20 1 7.5404253 0.9962383 0.0054372316038098075 1 +21 1 7.730527 0.99674207 0.0047078729886883648 1 +22 0 -5.183432 0.016643979 0.024214259971593023 0 23 1 ? ? ? 0 -24 0 -5.4686823 0.011754768 0.017059005370868869 0 -25 1 1.741828 0.73848027 0.43736871808798289 1 -26 0 -4.9710746 0.017053552 0.024815275746583213 0 -27 0 -4.0598474 0.033473443 0.049118721706209283 0 -28 0 -5.4818344 0.011639432 0.016890641442649377 0 -29 0 -5.8557806 0.00878995 0.012737279746493163 0 -30 0 -5.0985007 0.015506634 0.022546609667162744 0 -31 0 -4.9946866 0.016755886 0.024378450658658433 0 -32 1 7.46414 0.9954084 0.0066395103905325147 1 -33 0 -4.689259 0.021032775 0.030667534203555599 0 -34 0 -4.71424 0.020646105 0.030097814327393193 0 -35 0 -5.4818344 0.011639432 0.016890641442649377 0 -36 1 9.099108 0.99866724 0.0019240484689165348 1 -37 0 -1.113348 0.24456774 0.40462569514389818 0 -38 1 6.140953 0.9875705 0.018044317542393044 1 -39 1 2.5109034 0.83500063 0.26015080155533699 1 +24 0 -5.690696 0.011385534 0.016520077815512988 0 +25 1 1.6430082 0.75076216 0.41357214894050942 1 +26 0 -5.1501684 0.017062306 0.02482812490249775 0 +27 0 -4.190612 0.034712594 0.050969538544379397 0 +28 0 -5.679842 0.011478641 0.016655956359413658 0 +29 0 -6.1007133 0.00836595 0.012120283853048863 0 +30 0 -5.2624307 0.015690476 0.022816040551058049 0 +31 0 -5.194286 0.016509674 0.024017233951091451 0 +32 1 7.2332277 0.9952551 0.0068617173946663159 1 +33 0 -4.872356 0.020983735 0.030595266624822745 0 +34 0 -4.9028845 0.020512937 0.02990165675388249 0 +35 0 -5.679842 0.011478641 0.016655956359413658 0 +36 1 8.852162 0.99860686 0.0020112765691102027 1 +37 0 -1.2331572 0.25341713 0.42162569611996392 0 +38 1 6.101239 0.9888676 0.016150752045437752 1 +39 1 2.2340298 0.8251008 0.27735775097944682 1 40 0 ? ? ? 0 -41 1 3.3300762 0.90403634 0.14554732340521623 1 -42 1 8.577511 0.99802166 0.0028569650360322531 1 -43 1 0.49126053 0.5223421 0.93693314652059867 1 -44 1 8.255751 0.9974761 0.0036458194620463992 1 -45 0 -5.6322193 0.010397602 0.015079098451216091 0 -46 1 4.5673857 0.9601321 0.05869514874509995 1 -47 0 -5.95583 0.008152756 0.011810148833251818 0 -48 0 -3.4358397 0.052670274 0.078061439656217479 0 -49 1 5.3666544 0.9778563 0.032305655935016012 1 -50 1 2.5949678 0.84359986 0.2453692461908529 1 -51 1 0.12595749 0.45321622 1.141728585060551 1 -52 1 5.2992115 0.976721 0.033981595778575298 1 -53 1 8.407228 0.99774945 0.0032505195711204126 1 -54 1 7.649311 0.9960077 0.0057712269479890379 1 -55 1 4.478709 0.95747596 0.062691829736720397 1 -56 1 5.5541325 0.98073524 0.028064373404193169 1 -57 1 1.6657066 0.7271758 0.45962396180677062 1 -58 1 2.5265894 0.8366335 0.25733232225246744 1 -59 1 1.7368536 0.7377508 0.4387944595285721 1 -60 1 2.3288136 0.8150797 0.29498697868205404 1 -61 0 -5.5060835 0.011429708 0.016584543537116692 0 -62 1 6.380088 0.98961115 0.015066341261387473 1 -63 1 0.3348999 0.49270442 1.0212056771590654 1 -64 0 -5.95583 0.008152756 0.011810148833251818 0 -65 1 3.8072634 0.9311753 0.1028753176207461 1 -66 0 -4.046695 0.033797733 0.049602857843819056 0 -67 1 4.218014 0.9486548 0.076044929659653121 1 -68 1 10.826723 0.9996402 0.00051913703181331893 1 -69 0 -5.2716546 0.013623922 0.019790283751276339 0 -70 0 -3.4726496 0.051294286 0.075967458128490192 0 -71 1 7.895048 0.99668443 0.0047913008067115544 1 -72 0 -2.1755848 0.1263537 0.1948787725155719 0 -73 1 8.9055195 0.9984568 0.0022281210940979582 1 -74 1 2.5993576 0.84403867 0.244619004780572 1 -75 0 -4.0411606 0.033935104 0.049807988000729261 0 -76 0 -5.075033 0.01578076 0.022948375597840527 0 -77 0 -3.4995675 0.050309666 0.074470924297704952 0 -78 0 -3.6211967 0.04607984 0.068059573236894252 0 -79 0 -5.3911724 0.012457774 0.01808565768903711 0 -80 0 -2.7157316 0.087596945 0.13225681838671866 0 -81 0 -4.2284155 0.029574301 0.043310338334976231 0 -82 0 -3.4452734 0.052314345 0.077519494497433533 0 -83 0 -2.1223516 0.13087903 0.20237109943076298 0 -84 1 9.694054 0.99915093 0.0012254667568283894 1 -85 1 6.2895603 0.9888809 0.016131360230936344 1 -86 1 2.6168842 0.84578085 0.24164420028856959 1 -87 1 6.91914 0.9930738 0.010027129067340199 1 -88 0 -4.046695 0.033797733 0.049602857843819056 0 -89 0 -5.085745 0.01565505 0.022764118380358665 0 -90 0 -5.4686823 0.011754768 0.017059005370868869 0 -91 0 -5.189559 0.014486662 0.02105269812453239 0 -92 0 -4.046695 0.033797733 0.049602857843819056 0 -93 0 -5.95583 0.008152756 0.011810148833251818 0 -94 0 -4.9946866 0.016755886 0.024378450658658433 0 -95 0 -5.4686823 0.011754768 0.017059005370868869 0 -96 0 -5.663555 0.010155838 0.014726685232670428 0 -97 0 -3.5726995 0.047724236 0.070548678886274141 0 -98 1 8.590233 0.9980406 0.0028295658493644239 1 -99 1 10.917194 0.99966407 0.00048472853254951715 1 -100 1 4.8476696 0.9675202 0.047636346312542086 1 -101 1 -0.842803 0.28443503 1.813828922566693 0 -102 0 -3.7530966 0.04187613 0.06171590932041901 0 -103 1 1.7746449 0.74325943 0.42806223277958716 1 -104 1 12.140858 0.9998672 0.00019160139559387039 1 -105 1 2.5560703 0.83966714 0.25211056290937062 1 -106 1 9.259369 0.99881965 0.0017038920105930569 1 -107 1 6.720646 0.99195737 0.011649978820645172 1 -108 0 -5.5617743 0.0109620895 0.015902273389359543 0 -109 1 6.871727 0.992822 0.010393022735684912 1 -110 0 -2.766693 0.08455608 0.12745658613888911 0 -111 1 3.848031 0.93313104 0.099848402613633994 1 -112 1 9.42577 0.9989595 0.0015019320975399703 1 -113 1 9.506622 0.9990213 0.0014126689718023965 1 -114 0 -3.0727453 0.06823268 0.10195836010283554 0 -115 0 -4.643991 0.02175159 0.031727234757411307 0 -116 0 -0.6618881 0.31317252 0.54198033042993166 0 -117 1 9.617277 0.9991 0.0012989676266954972 1 -118 0 -5.3621607 0.012731452 0.018485527093437829 0 -119 0 -3.9435177 0.036449015 0.053567088295960068 0 -120 0 -4.8696556 0.018392263 0.026781474537826103 0 -121 0 -3.469522 0.051409863 0.076143226767382041 0 -122 1 9.680523 0.99914217 0.0012381182790972595 1 -123 1 3.8165932 0.9316275 0.10217485034891999 1 -124 1 7.6522446 0.99601656 0.0057583629403995997 1 -125 0 -5.95583 0.008152756 0.011810148833251818 0 -126 1 8.564951 0.9980027 0.0028843647430662487 1 -127 0 -4.520691 0.023833439 0.034800762077037174 0 -128 1 4.848981 0.9675514 0.04758977495588574 1 -129 0 -5.717684 0.009751258 0.014137132147530099 0 -130 0 -3.4726496 0.051294286 0.075967458128490192 0 -131 0 -4.9946866 0.016755886 0.024378450658658433 0 -132 1 8.60223 0.9980583 0.0028039765128859018 1 -133 0 -4.810811 0.019215731 0.027992255281483396 0 -134 0 -4.917177 0.01775261 0.025841665902698104 0 -135 0 -2.7288966 0.08680205 0.13100047471163517 0 -136 0 -4.533843 0.02360242 0.034459376167277234 0 -137 0 -5.4949865 0.011525216 0.016723931014560982 0 -138 0 -4.2402444 0.02931786 0.042929146807739925 0 +41 1 3.2122545 0.90836364 0.13865813542624633 1 +42 1 8.198594 0.99771404 0.0033017145416675006 1 +43 1 0.22008419 0.5056527 0.98378119004527576 1 +44 1 7.915265 0.99716717 0.0040927094395791829 1 +45 0 -5.8607707 0.010020453 0.014529374846107383 0 +46 1 4.2778234 0.9570013 0.063407167487105412 1 +47 0 -6.1762524 0.007903442 0.011447553633965282 0 +48 0 -3.5576057 0.054948628 0.081535339956679373 0 +49 1 5.1043253 0.9765691 0.034205941253008319 1 +50 1 2.3847685 0.84100354 0.24981622656094413 1 +51 1 -0.002776146 0.4634279 1.1095831887743572 0 +52 1 5.171505 0.97770804 0.032524375845745371 1 +53 1 8.174706 0.9976723 0.0033620476470297487 1 +54 1 7.5763617 0.99633914 0.0052911922877945505 1 +55 1 4.3368845 0.9588088 0.060684975225045172 1 +56 1 5.228385 0.9786298 0.031164839778719866 1 +57 1 1.4711056 0.7255575 0.46283813124317547 1 +58 1 2.3726006 0.8397646 0.25194313030203674 1 +59 1 1.6023321 0.7449402 0.42480343559488037 1 +60 1 2.1442842 0.81505156 0.29503677578365628 1 +61 0 -5.7187867 0.011148015 0.016173505693764696 0 +62 1 6.0357714 0.988307 0.016968837049659936 1 +63 1 0.09457874 0.48184475 1.0533596997094408 1 +64 0 -6.1762524 0.007903442 0.011447553633965282 0 +65 1 3.6413336 0.9321066 0.10143311548516794 1 +66 0 -4.2014656 0.034437597 0.050558592955505986 0 +67 1 3.980445 0.9466932 0.079031162901014043 1 +68 1 10.612863 0.99963355 0.00052877155866913161 1 +69 0 -5.4978714 0.013156473 0.019106744037250722 0 +70 0 -3.6295166 0.052182022 0.077318069673474266 0 +71 1 7.5565577 0.9962839 0.0053712013587022198 1 +72 0 -2.4480972 0.11892297 0.1826599415467918 0 +73 1 8.905442 0.998662 0.0019316258242218953 1 +74 1 2.4551315 0.8480158 0.23783697509565524 1 +75 0 -4.274302 0.032645807 0.047883870748970089 0 +76 0 -5.334609 0.014866402 0.021608707816189145 0 +77 0 -3.644158 0.051635087 0.076485806285536237 0 +78 0 -3.7829442 0.04671344 0.069018140747015844 0 +79 0 -5.6008434 0.01217936 0.017678981190010004 0 +80 0 -2.9382496 0.08512094 0.12834705009270028 0 +81 0 -4.3650923 0.030538203 0.044744047143781454 0 +82 0 -3.5927935 0.05357801 0.079444502554735716 0 +83 0 -2.2981849 0.13137244 0.203190362054017 0 +84 1 9.3922205 0.99907494 0.0013352030401460008 1 +85 1 5.7727385 0.98575985 0.020691867334853394 1 +86 1 2.3984623 0.84238845 0.24744243731746635 1 +87 1 6.6674128 0.9927282 0.010529358093716759 1 +88 0 -4.2014656 0.034437597 0.050558592955505986 0 +89 0 -5.2789135 0.015498414 0.022534564190655703 0 +90 0 -5.690696 0.011385534 0.016520077815512988 0 +91 0 -5.3470583 0.014728642 0.021406976586218332 0 +92 0 -4.2014656 0.034437597 0.050558592955505986 0 +93 0 -6.1762524 0.007903442 0.011447553633965282 0 +94 0 -5.194286 0.016509674 0.024017233951091451 0 +95 0 -5.690696 0.011385534 0.016520077815512988 0 +96 0 -5.8434687 0.010151575 0.014720472426273983 0 +97 0 -3.7050557 0.049417876 0.073116822887907204 0 +98 1 8.223899 0.99775743 0.0032389707884948513 1 +99 1 10.822373 0.9996874 0.00045109501765938913 1 +100 1 4.636462 0.9669095 0.048547189602392155 1 +101 1 -1.0467243 0.28111085 1.830788941401543 0 +102 0 -3.9209185 0.04226496 0.062301510291738804 0 +103 1 1.630683 0.74900746 0.41694800045412517 1 +104 1 11.827565 0.9998542 0.00021035011616098864 1 +105 1 2.3993664 0.8424796 0.24728636487289407 1 +106 1 9.16854 0.99890393 0.0015821617216926738 1 +107 1 6.4510603 0.9914414 0.012400549942853872 1 +108 0 -5.7906036 0.010562822 0.01531998557959283 0 +109 1 6.7095194 0.9929553 0.010199282386765415 1 +110 0 -3.014533 0.08071872 0.1214217302397668 0 +111 1 3.7277098 0.93614 0.095203792231295992 1 +112 1 9.536014 0.9991705 0.0011972379098122031 1 +113 1 9.218173 0.9989444 0.0015237107281484779 1 +114 0 -3.277872 0.067075275 0.10016741582961504 0 +115 0 -4.8319454 0.02162322 0.031537930802890933 0 +116 0 -0.7799158 0.3237837 0.56444329330798382 0 +117 1 9.464227 0.9991241 0.0012641962566632629 1 +118 0 -5.5888796 0.0122891 0.017839264252685727 0 +119 0 -4.1281476 0.03633683 0.053399129355539122 0 +120 0 -5.088218 0.017869117 0.026012797384040408 0 +121 0 -3.6317377 0.0520987 0.077191246572681393 0 +122 1 9.371118 0.99906003 0.0013567209372589222 1 +123 1 3.5896034 0.92957914 0.1053504021682207 1 +124 1 7.4270334 0.9959016 0.0059249132248314022 1 +125 0 -6.1762524 0.007903442 0.011447553633965282 0 +126 1 8.480126 0.9981531 0.0026669908139932258 1 +127 0 -4.697876 0.023884254 0.034875864219680933 0 +128 1 4.7593575 0.96976763 0.044288996105927125 1 +129 0 -6.02982 0.008824386 0.012787401320266147 0 +130 0 -3.6295166 0.052182022 0.077318069673474266 0 +131 0 -5.194286 0.016509674 0.024017233951091451 0 +132 1 8.352777 0.997966 0.0029374423855024019 1 +133 0 -4.9906077 0.01921693 0.027994019763783125 0 +134 0 -5.1044335 0.017654385 0.025697403625828272 0 +135 0 -2.8619413 0.08974136 0.13565156009217885 0 +136 0 -4.687022 0.02407708 0.035160889311831381 0 +137 0 -5.668988 0.0115725 0.016792944693470865 0 +138 0 -4.4064746 0.029621843 0.04338101918277134 0 139 0 ? ? ? 0 -140 0 -5.4949865 0.011525216 0.016723931014560982 0 -141 0 -5.9689827 0.008072471 0.011693375209389432 0 -142 1 4.4324036 0.95602256 0.064883430994521443 1 -143 0 -4.643991 0.02175159 0.031727234757411307 0 -144 0 -5.4818344 0.011639432 0.016890641442649377 0 +140 0 -5.668988 0.0115725 0.016792944693470865 0 +141 0 -6.1653986 0.007968302 0.011541875502083741 0 +142 1 4.307641 0.9579231 0.062018229186656552 1 +143 0 -4.8319454 0.02162322 0.031537930802890933 0 +144 0 -5.679842 0.011478641 0.016655956359413658 0 145 0 ? ? ? 0 -146 1 1.3394356 0.6754276 0.56612692807369991 1 -147 0 -5.4154215 0.012233487 0.017758035691552793 0 -148 0 -1.012373 0.25899473 0.43244428867240287 0 -149 1 11.461615 0.99977773 0.00032069729386163213 1 -150 0 -5.559344 0.010982096 0.015931457121809788 0 -151 1 5.006485 0.9711001 0.042308092038938809 1 -152 1 9.715748 0.99916476 0.0012054999542047689 1 -153 0 -4.1214976 0.03199297 0.046910567277187794 0 -154 0 -6.442978 0.0056482 0.0081717296030013944 0 -155 1 3.7769232 0.9296856 0.10518519653096636 1 -156 0 -5.5348053 0.0111861285 0.016229112739800475 0 -157 0 -4.9946866 0.016755886 0.024378450658658433 0 +146 1 1.0868425 0.6638557 0.59105847393586752 1 +147 0 -5.639788 0.011828791 0.017167072843240438 0 +148 0 -1.259141 0.24970378 0.414467804588733 0 +149 1 11.153153 0.9997568 0.00035088727509285797 1 +150 0 -5.7696943 0.010729991 0.01556375426175181 0 +151 1 4.855405 0.9718334 0.041219067296589307 1 +152 1 9.511897 0.99915516 0.0012193561973709755 1 +153 0 -4.257647 0.033047408 0.048482936073668756 0 +154 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +155 1 3.6720238 0.9335661 0.0991759307008305 1 +156 0 -5.7834425 0.010619783 0.015403042059905666 0 +157 0 -5.194286 0.016509674 0.024017233951091451 0 158 0 ? ? ? 0 -159 1 12.346203 0.99988633 0.0001639947780942108 1 -160 1 9.039494 0.99860567 0.002012998795947356 1 -161 0 -3.8496675 0.039033547 0.057442027050147004 0 -162 0 -4.520691 0.023833439 0.034800762077037174 0 -163 0 -3.387055 0.054547653 0.080923349462577995 0 +159 1 12.03571 0.99987555 0.00017956106132345373 1 +160 1 8.89737 0.99865377 0.0019435085751829124 1 +161 0 -4.0086412 0.03965031 0.058368267792054167 0 +162 0 -4.697876 0.023884254 0.034875864219680933 0 +163 0 -3.6110554 0.052879456 0.078380040321185507 0 164 0 ? ? ? 0 -165 0 -3.3999205 0.05404651 0.080158845547395108 0 -166 1 7.976183 0.9968817 0.004505750926832852 1 -167 1 8.355644 0.99765986 0.0033800618776249807 1 -168 0 -4.520691 0.023833439 0.034800762077037174 0 -169 0 -6.2282124 0.00664095 0.0096128205240473656 0 -170 0 -5.4949865 0.011525216 0.016723931014560982 0 -171 0 -5.4686823 0.011754768 0.017059005370868869 0 -172 0 -5.95583 0.008152756 0.011810148833251818 0 -173 1 15.1560135 0.9999865 1.9434170443242565E-05 1 -174 1 6.1769247 0.9879011 0.017561488134486943 1 -175 1 7.842922 0.99655116 0.004984229936716684 1 -176 0 -4.9946866 0.016755886 0.024378450658658433 0 -177 1 4.766121 0.96551895 0.050623517901417933 1 -178 0 -4.046695 0.033797733 0.049602857843819056 0 -179 1 2.290575 0.8106676 0.30281765649168213 1 -180 0 -5.559344 0.010982096 0.015931457121809788 0 -181 0 -6.442978 0.0056482 0.0081717296030013944 0 -182 0 -3.0987039 0.06699134 0.10003761775121778 0 -183 1 9.159964 0.9987273 0.0018372561468384253 1 -184 1 6.2014647 0.98812157 0.017239546569330137 1 -185 0 -5.0853486 0.015659682 0.022770907823479652 0 -186 1 5.7654104 0.9835412 0.023942621168576837 1 -187 1 13.977451 0.99996704 4.7553986690113E-05 1 -188 1 9.065283 0.99863267 0.0019739908621602127 1 -189 0 -4.7540584 0.020044118 0.029211295141143595 0 -190 1 11.957218 0.99984735 0.0002202405944450654 1 -191 1 10.956873 0.999674 0.00047036322693564959 1 -192 0 -4.0598474 0.033473443 0.049118721706209283 0 -193 0 -5.4686823 0.011754768 0.017059005370868869 0 -194 0 -4.520691 0.023833439 0.034800762077037174 0 -195 0 -4.046695 0.033797733 0.049602857843819056 0 -196 0 6.8652763 0.992787 7.1151855643261381 1 -197 0 -2.6564164 0.09126051 0.13806131755901335 0 -198 0 -6.442978 0.0056482 0.0081717296030013944 0 -199 0 -5.0078387 0.016592305 0.02413845106369562 0 -200 1 10.36586 0.9994898 0.00073627359086373204 1 -201 1 9.869495 0.9992566 0.0010728826448553252 1 -202 0 -5.4686823 0.011754768 0.017059005370868869 0 -203 0 -3.5726995 0.047724236 0.070548678886274141 0 -204 0 -5.4686823 0.011754768 0.017059005370868869 0 -205 1 12.086601 0.99986166 0.00019959967319162043 1 -206 1 5.944169 0.98559827 0.020928376857247459 1 -207 0 -5.559344 0.010982096 0.015931457121809788 0 -208 0 -5.559344 0.010982096 0.015931457121809788 0 -209 0 -3.6633615 0.04469411 0.065965336354842946 0 -210 1 14.534113 0.99997836 3.1215188826316377E-05 1 -211 1 9.64962 0.99912184 0.0012674667953664828 1 -212 0 -5.4686823 0.011754768 0.017059005370868869 0 -213 1 14.529058 0.9999783 3.1301182014901005E-05 1 -214 1 13.868914 0.9999642 5.1681712271066226E-05 1 -215 1 7.643734 0.9959909 0.0057955739072098177 1 -216 0 -5.95583 0.008152756 0.011810148833251818 0 -217 0 -5.4686823 0.011754768 0.017059005370868869 0 -218 1 7.8867817 0.99666363 0.0048214119279433977 1 -219 0 -2.511506 0.100795746 0.15327923389796996 0 -220 0 -5.1632547 0.014774317 0.021473858071037247 0 -221 1 10.395218 0.999501 0.00072009905980506843 1 -222 1 -2.214662 0.1231175 3.0218922639873447 0 -223 1 5.7424126 0.9832564 0.024360423750620218 1 -224 1 9.995327 0.99932426 0.00097521318777220637 1 -225 0 -5.95583 0.008152756 0.011810148833251818 0 -226 1 10.225868 0.9994326 0.00081878372812222485 1 -227 1 7.459608 0.9953926 0.0066624033876966004 1 -228 0 -5.559344 0.010982096 0.015931457121809788 0 -229 1 12.666515 0.9999109 0.00012856275965089447 1 -230 1 6.1583214 0.9877312 0.017809586200967932 1 -231 1 8.623034 0.99808866 0.0027601224428665879 1 -232 0 1.2822819 0.6658516 1.5814390985815938 1 -233 1 6.3825197 0.9896301 0.015038709216571163 1 -234 0 -2.8964381 0.07724253 0.115976584697302 0 +165 0 -3.5511756 0.055202637 0.081923156614349499 0 +166 1 7.750617 0.99679124 0.0046367000185424355 1 +167 1 8.140458 0.99761117 0.0034504833010182072 1 +168 0 -4.697876 0.023884254 0.034875864219680933 0 +169 0 -6.418877 0.006582828 0.0095284094443157507 0 +170 0 -5.668988 0.0115725 0.016792944693470865 0 +171 0 -5.690696 0.011385534 0.016520077815512988 0 +172 0 -6.1762524 0.007903442 0.011447553633965282 0 +173 1 14.842399 0.9999852 2.1326006327376515E-05 1 +174 1 6.0863256 0.98874223 0.016333639242311616 1 +175 1 7.5728846 0.9963295 0.0053051741355733617 1 +176 0 -5.194286 0.016509674 0.024017233951091451 0 +177 1 4.2855787 0.9572429 0.063043031252521448 1 +178 0 -4.2014656 0.034437597 0.050558592955505986 0 +179 1 2.15761 0.8165715 0.29234892522890787 1 +180 0 -5.7696943 0.010729991 0.01556375426175181 0 +181 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +182 0 -3.2086458 0.07043917 0.10537881324670834 0 +183 1 8.944306 0.9987008 0.0018755715579931295 1 +184 1 6.0725775 0.9886255 0.016503937371878118 1 +185 0 -5.2732844 0.015563743 0.022630301619727688 0 +186 1 5.448736 0.98186153 0.026408513896858703 1 +187 1 13.897843 0.9999697 4.3684254683749625E-05 1 +188 1 8.907805 0.9986644 0.001928181566876736 1 +189 0 -5.0173383 0.018838223 0.027437062895142233 0 +190 1 11.761691 0.99984676 0.00022110063923937645 1 +191 1 10.76023 0.9996723 0.00047285779070154501 1 +192 0 -4.190612 0.034712594 0.050969538544379397 0 +193 0 -5.690696 0.011385534 0.016520077815512988 0 +194 0 -4.697876 0.023884254 0.034875864219680933 0 +195 0 -4.2014656 0.034437597 0.050558592955505986 0 +196 0 6.7605906 0.99322134 7.204784719513877 1 +197 0 -2.7392664 0.09764401 0.14823138694212687 0 +198 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +199 0 -5.183432 0.016643979 0.024214259971593023 0 +200 1 10.20231 0.9994996 0.00072207784908216624 1 +201 1 9.669415 0.99925035 0.0010819184794421913 1 +202 0 -5.690696 0.011385534 0.016520077815512988 0 +203 0 -3.7050557 0.049417876 0.073116822887907204 0 +204 0 -5.690696 0.011385534 0.016520077815512988 0 +205 1 11.778428 0.99984866 0.00021834849770230094 1 +206 1 5.671749 0.9846428 0.022327638171871653 1 +207 0 -5.7696943 0.010729991 0.01556375426175181 0 +208 0 -5.7696943 0.010729991 0.01556375426175181 0 +209 0 -3.7840543 0.046675935 0.06896138029379914 0 +210 1 14.311627 0.9999779 3.190313447851316E-05 1 +211 1 9.471307 0.99912876 0.0012574830688691823 1 +212 0 -5.690696 0.011385534 0.016520077815512988 0 +213 1 14.345925 0.9999784 3.1129195642857458E-05 1 +214 1 13.768859 0.9999666 4.8155945935111852E-05 1 +215 1 7.442033 0.9959478 0.0058579972436691276 1 +216 0 -6.1762524 0.007903442 0.011447553633965282 0 +217 0 -5.690696 0.011385534 0.016520077815512988 0 +218 1 7.5836535 0.9963593 0.005262020720692949 1 +219 0 -2.6475506 0.10395207 0.15835219426781896 0 +220 0 -5.368766 0.014491433 0.021059682660687775 0 +221 1 10.2217455 0.99950695 0.00071149565971861516 1 +222 1 -2.2922978 0.13188319 2.9226674163744848 0 +223 1 5.585368 0.98361915 0.023828266836395273 1 +224 1 9.954375 0.999396 0.00087161334294032168 1 +225 0 -6.1762524 0.007903442 0.011447553633965282 0 +226 1 10.120683 0.9994676 0.00076827904844171643 1 +227 1 7.2224054 0.9952162 0.0069181385401670542 1 +228 0 -5.7696943 0.010729991 0.01556375426175181 0 +229 1 12.354081 0.99990225 0.00014103266690546063 1 +230 1 6.003766 0.9880229 0.017383580210835827 1 +231 1 8.352307 0.9979653 0.0029384763849410163 1 +232 0 1.1177626 0.6690727 1.5954137285901675 1 +233 1 6.363435 0.99085826 0.013249401565498231 1 +234 0 -3.1496835 0.07342657 0.11002277734958922 0 235 0 ? ? ? 0 -236 1 11.420414 0.99977064 0.00033093257283884215 1 -237 1 6.535795 0.9907579 0.01339555438417007 1 -238 1 12.422876 0.9998928 0.00015470668910224405 1 -239 1 5.9025297 0.985143 0.021594929859074165 1 -240 0 -2.0179915 0.14015017 0.21784338265744002 0 -241 0 -4.0004973 0.034961022 0.051340881416726061 0 -242 0 -4.9946866 0.016755886 0.024378450658658433 0 -243 0 -2.6953988 0.088837564 0.13421982380863931 0 -244 0 -5.4686823 0.011754768 0.017059005370868869 0 -245 0 -2.817525 0.08161871 0.12283484643891715 0 -246 1 11.424002 0.9997713 0.00032998645156814019 1 -247 1 3.104393 0.8881221 0.17117009101578401 1 -248 0 -3.0615559 0.068774305 0.10279722872890984 0 +236 1 11.114438 0.99974954 0.00036138080683877994 1 +237 1 6.258853 0.9901104 0.014338700186334048 1 +238 1 12.123448 0.99988353 0.00016803683549596564 1 +239 1 5.86524 0.9867126 0.019298199740636904 1 +240 0 -2.1425428 0.14545055 0.22676411200116661 0 +241 0 -4.1330147 0.03620769 0.053205806129218658 0 +242 0 -5.194286 0.016509674 0.024017233951091451 0 +243 0 -2.8897204 0.08803374 0.13294764994008823 0 +244 0 -5.690696 0.011385534 0.016520077815512988 0 +245 0 -2.9256601 0.085868075 0.1295257095799256 0 +246 1 11.219155 0.9997687 0.00033377094037375917 1 +247 1 3.0079794 0.8946139 0.16066288569750423 1 +248 0 -3.2373443 0.06902611 0.10318739217744149 0 249 0 ? ? ? 0 -250 0 -6.021953 0.0077569974 0.011234611610460798 0 -251 1 8.872498 0.9984177 0.0022846196981478615 1 -252 0 4.5387735 0.95929295 4.6185774530404009 1 -253 1 8.577511 0.99802166 0.0028569650360322531 1 -254 1 6.380088 0.98961115 0.015066341261387473 1 -255 1 4.052039 0.94216394 0.085949972449708326 1 -256 0 -5.4949865 0.011525216 0.016723931014560982 0 -257 0 -5.0078387 0.016592305 0.02413845106369562 0 -258 0 -4.520691 0.023833439 0.034800762077037174 0 -259 0 2.9647484 0.8771559 3.0250995258103677 1 -260 1 9.870926 0.99925745 0.0010716778711863072 1 -261 1 12.206299 0.99987364 0.00018231312886902128 1 -262 1 9.653839 0.99912465 0.0012634216564773898 1 -263 1 8.981979 0.9985436 0.002102643541396716 1 -264 1 5.664708 0.9822578 0.025826400239286989 1 -265 0 -2.4948754 0.10194495 0.15512421639444568 0 -266 1 7.3661633 0.99505585 0.0071505858244227481 1 -267 1 3.3009605 0.902103 0.14863591753511315 1 -268 1 9.372967 0.998917 0.0015633090806578933 1 -269 0 -5.4686823 0.011754768 0.017059005370868869 0 -270 1 6.031377 0.9865077 0.019597763036488004 1 -271 0 -3.5726995 0.047724236 0.070548678886274141 0 -272 1 3.3009605 0.902103 0.14863591753511315 1 -273 1 0.21747208 0.4704687 1.0878293505835543 1 -274 0 -4.323663 0.02756967 0.040333203653413245 0 +250 0 -6.268999 0.0073701376 0.010672237396626542 0 +251 1 8.957709 0.99871397 0.0018565428782942226 1 +252 0 4.4631453 0.9624317 4.7343414630977234 1 +253 1 8.198594 0.99771404 0.0033017145416675006 1 +254 1 6.0357714 0.988307 0.016968837049659936 1 +255 1 3.8202343 0.94021213 0.088941800692148978 1 +256 0 -5.668988 0.0115725 0.016792944693470865 0 +257 0 -5.183432 0.016643979 0.024214259971593023 0 +258 0 -4.697876 0.023884254 0.034875864219680933 0 +259 0 2.6695328 0.8678251 2.9194797601837759 1 +260 1 9.885512 0.99936366 0.0009183356077711259 1 +261 1 11.814006 0.9998527 0.00021250021436790644 1 +262 1 9.426362 0.9990986 0.0013010332790095319 1 +263 1 9.02839 0.998781 0.0017596813178296863 1 +264 1 5.4595184 0.9820067 0.026195185056672989 1 +265 0 -2.6705885 0.10233448 0.15575010895131483 0 +266 1 7.2168283 0.995196 0.006947430020272323 1 +267 1 3.1091843 0.90164 0.14937657807139287 1 +268 1 9.093087 0.9988394 0.0016753955271333627 1 +269 0 -5.690696 0.011385534 0.016520077815512988 0 +270 1 5.858514 0.9866455 0.019396246052056781 1 +271 0 -3.7050557 0.049417876 0.073116822887907204 0 +272 1 3.1091843 0.90164 0.14937657807139287 1 +273 1 0.07871342 0.47883877 1.0623881228399479 1 +274 0 -4.5050516 0.027545111 0.04029676848184504 0 275 0 ? ? ? 0 -276 0 -5.0078387 0.016592305 0.02413845106369562 0 -277 0 -5.95583 0.008152756 0.011810148833251818 0 -278 0 -5.4686823 0.011754768 0.017059005370868869 0 -279 1 7.127905 0.9940822 0.0085629246022787941 1 -280 0 -4.520691 0.023833439 0.034800762077037174 0 -281 0 -4.689259 0.021032775 0.030667534203555599 0 -282 1 4.4381237 0.9562046 0.064608759119747658 1 -283 1 6.0636253 0.9868295 0.019127222915579972 1 -284 1 7.431343 0.99529326 0.006806421628886848 1 -285 1 14.218481 0.9999725 3.9642545670177728E-05 1 -286 1 15.281263 0.9999877 1.7714321792245208E-05 1 -287 0 -4.917177 0.01775261 0.025841665902698104 0 -288 1 2.2163515 0.80187416 0.31855224459431192 1 -289 1 8.312019 0.9975813 0.0034936687628036997 1 -290 0 -6.442978 0.0056482 0.0081717296030013944 0 -291 0 -5.4686823 0.011754768 0.017059005370868869 0 +276 0 -5.183432 0.016643979 0.024214259971593023 0 +277 0 -6.1762524 0.007903442 0.011447553633965282 0 +278 0 -5.690696 0.011385534 0.016520077815512988 0 +279 1 6.9710016 0.9942162 0.0083684784383695216 1 +280 0 -4.697876 0.023884254 0.034875864219680933 0 +281 0 -4.872356 0.020983735 0.030595266624822745 0 +282 1 4.3149786 0.958147 0.061681098077334774 1 +283 1 5.990473 0.98790294 0.01755878975848741 1 +284 1 7.4265547 0.9959001 0.0059270718565448965 1 +285 1 13.994821 0.99997187 4.0588476552067527E-05 1 +286 1 15.019143 0.99998707 1.8660238296575832E-05 1 +287 0 -5.1044335 0.017654385 0.025697403625828272 0 +288 1 2.0468407 0.8036413 0.31537635264229774 1 +289 1 8.113109 0.9975612 0.0035227183934827171 1 +290 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +291 0 -5.690696 0.011385534 0.016520077815512988 0 292 1 ? ? ? 0 -293 1 5.542122 0.9805624 0.028318669172894068 1 +293 1 5.5823574 0.9835823 0.023882295505951855 1 294 0 ? ? ? 0 -295 1 7.7866364 0.99640125 0.0052012629006476284 1 -296 0 1.823431 0.75025773 2.0014880731401443 1 +295 1 7.634673 0.9964971 0.0050624961123187946 1 +296 0 1.6353331 0.74967045 1.9980994682277313 1 297 0 ? ? ? 0 -298 0 -2.725597 0.087000675 0.13131430208208369 0 -299 1 7.8274364 0.99651057 0.0050429938896844372 1 -300 1 7.348074 0.9949879 0.0072491063826461647 1 -301 0 -5.4686823 0.011754768 0.017059005370868869 0 -302 1 15.735764 0.9999913 1.2554788140693439E-05 1 -303 0 -5.4686823 0.011754768 0.017059005370868869 0 -304 1 5.9607973 0.98577625 0.020667878309752356 1 -305 1 8.459469 0.9978367 0.003124349823870973 1 -306 0 -5.4686823 0.011754768 0.017059005370868869 0 -307 0 -5.4686823 0.011754768 0.017059005370868869 0 -308 1 7.4225903 0.9952621 0.0068516084791740307 1 -309 0 -1.7474074 0.16675761 0.26319186454303012 0 -310 0 -5.3911724 0.012457774 0.01808565768903711 0 -311 0 -6.442978 0.0056482 0.0081717296030013944 0 -312 1 3.6294699 0.9220128 0.11714130634122567 1 -313 0 -6.442978 0.0056482 0.0081717296030013944 0 -314 0 -6.0464916 0.00761503 0.011028209313642047 0 +298 0 -2.8376803 0.091257066 0.13805585285798452 0 +299 1 7.8606234 0.99704754 0.0042657947038420088 1 +300 1 7.1416483 0.9949153 0.0073543752558372791 1 +301 0 -5.690696 0.011385534 0.016520077815512988 0 +302 1 15.431024 0.9999905 1.3672685532662985E-05 1 +303 0 -5.690696 0.011385534 0.016520077815512988 0 +304 1 6.007621 0.9880575 0.017333101527569443 1 +305 1 8.327315 0.9979264 0.0029946581351453267 1 +306 0 -5.690696 0.011385534 0.016520077815512988 0 +307 0 -5.690696 0.011385534 0.016520077815512988 0 +308 1 7.1634607 0.99499834 0.007233982175612351 1 +309 0 -1.8969002 0.17018971 0.2691465448779084 0 +310 0 -5.6008434 0.01217936 0.017678981190010004 0 +311 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +312 1 3.4877834 0.92434883 0.11349069558673246 1 +313 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +314 0 -6.255251 0.007446876 0.010783773513399939 0 315 0 ? ? ? 0 -316 1 3.6177397 0.9213705 0.11814667788191786 1 -317 1 9.215706 0.99877995 0.0017612310516048875 1 -318 0 -5.1966968 0.014409562 0.020939834998732896 0 -319 0 2.6369457 0.84775543 2.7155373416013662 1 -320 1 7.3824844 0.9951164 0.0070627872070711771 1 +316 1 3.4135065 0.92031074 0.11980703679753058 1 +317 1 9.002616 0.998757 0.0017943785337299681 1 +318 0 -5.4537 0.013598964 0.019753779349788943 0 +319 0 2.4164848 0.84419626 2.6821982295870535 1 +320 1 7.184394 0.99507684 0.0071201668008942769 1 321 0 ? ? ? 0 -322 0 -4.520691 0.023833439 0.034800762077037174 0 -323 1 5.5612926 0.98083764 0.027913746222999549 1 -324 0 -5.4686823 0.011754768 0.017059005370868869 0 -325 0 -4.1927576 0.030360568 0.044479725758863976 0 -326 1 4.4103804 0.95531476 0.065951946768905437 1 -327 0 -5.95583 0.008152756 0.011810148833251818 0 -328 1 3.373887 0.9068811 0.141014691181642 1 -329 1 7.8321342 0.9965229 0.0050251314657524084 1 -330 1 5.8562517 0.98462033 0.022360562903046506 1 -331 0 -3.2490892 0.060203623 0.089579888898278004 0 -332 0 -3.1363668 0.06522752 0.097312838862202627 0 -333 1 4.914962 0.9690866 0.04530251936714582 1 -334 1 5.9119453 0.98524714 0.021442445493118613 1 -335 0 -6.442978 0.0056482 0.0081717296030013944 0 -336 1 5.54352 0.9805826 0.028288940560717461 1 -337 0 -5.4686823 0.011754768 0.017059005370868869 0 -338 0 -6.0464916 0.00761503 0.011028209313642047 0 -339 1 5.684024 0.98251134 0.025454033740074194 1 -340 1 6.620782 0.9913299 0.012562837713558307 1 -341 0 -5.4686823 0.011754768 0.017059005370868869 0 -342 0 -5.9689827 0.008072471 0.011693375209389432 0 -343 0 -6.442978 0.0056482 0.0081717296030013944 0 -344 1 10.162451 0.99940467 0.00085913711935366719 1 -345 0 -6.442978 0.0056482 0.0081717296030013944 0 -346 0 -3.335825 0.0565869 0.084038459237484736 0 -347 0 -6.139584 0.0070995055 0.01027895246563207 0 -348 1 0.15727425 0.45910957 1.1230895756645052 1 -349 1 4.0622606 0.94258505 0.085305293794040227 1 -350 0 -3.93614 0.03664608 0.053862176491974861 0 -351 0 -4.9946866 0.016755886 0.024378450658658433 0 -352 0 0.4719286 0.51868206 1.0549379053712531 1 -353 1 8.696344 0.9981919 0.0026109079692710987 1 -354 0 -5.95583 0.008152756 0.011810148833251818 0 -355 0 -4.246154 0.029190555 0.042739950676879641 0 -356 1 -0.69921684 0.30711415 1.7031530909424435 0 -357 1 12.852016 0.9999226 0.00011170705633068974 1 -358 1 5.5822067 0.9811335 0.027478611445726227 1 -359 1 5.3672857 0.97786665 0.032290354697391166 1 -360 1 15.333874 0.9999882 1.7026382905914664E-05 1 -361 1 6.31769 0.9891131 0.0157926107591757 1 -362 0 -3.5059962 0.050077174 0.074117784460681763 0 -363 0 -2.065846 0.13583243 0.21061700181839857 0 -364 0 -4.9946866 0.016755886 0.024378450658658433 0 -365 0 -5.4818344 0.011639432 0.016890641442649377 0 -366 1 13.694569 0.9999591 5.899125529457858E-05 1 -367 1 11.299244 0.9997486 0.00036275701338483524 1 -368 0 -5.8557806 0.00878995 0.012737279746493163 0 -369 0 -5.4592943 0.011837784 0.017180202147298197 0 -370 0 -3.8947306 0.03777132 0.055548296888384752 0 -371 0 -5.8557806 0.00878995 0.012737279746493163 0 -372 0 -4.2402444 0.02931786 0.042929146807739925 0 -373 0 -3.7544198 0.041835874 0.061655295893265259 0 -374 0 -4.71424 0.020646105 0.030097814327393193 0 -375 0 -6.442978 0.0056482 0.0081717296030013944 0 -376 0 -5.95583 0.008152756 0.011810148833251818 0 -377 0 -6.0464916 0.00761503 0.011028209313642047 0 -378 0 -3.8038664 0.040357865 0.059431592056106 0 -379 0 -2.2557268 0.11979375 0.18408647967245978 0 -380 0 -6.442978 0.0056482 0.0081717296030013944 0 -381 1 10.076408 0.99936455 0.00091704491714961699 1 -382 0 -3.5972104 0.046886295 0.069279759305546196 0 -383 0 -5.9689827 0.008072471 0.011693375209389432 0 -384 0 -5.9689827 0.008072471 0.011693375209389432 0 -385 0 -3.7061968 0.043327074 0.063902325254427392 0 -386 1 6.0875874 0.9870637 0.018784894450298191 1 -387 0 -2.33456 0.113630086 0.17401918227784446 0 -388 0 -5.2848067 0.013490497 0.01959514665447612 0 -389 0 -3.322411 0.057132576 0.084873165657017816 0 -390 0 -5.6504025 0.0102566285 0.014873594611876085 0 -391 1 10.030338 0.99934196 0.00094965672085389622 1 -392 0 -5.0078387 0.016592305 0.02413845106369562 0 -393 0 -6.53364 0.0052747848 0.007630046968004383 0 -394 0 -5.241206 0.013937827 0.020249481181878783 0 -395 0 -5.0078387 0.016592305 0.02413845106369562 0 -396 0 -4.520691 0.023833439 0.034800762077037174 0 -397 0 -5.0209913 0.016430289 0.023900786364877163 0 -398 0 -4.683385 0.021124722 0.030803042926351529 0 -399 0 -5.7283545 0.009673405 0.014023712470766305 0 -400 1 10.056744 0.999355 0.00093081234331069047 1 -401 0 -5.4949865 0.011525216 0.016723931014560982 0 -402 0 -3.2177973 0.061560776 0.091664779980531616 0 -403 0 -4.145746 0.03142818 0.046069066605854456 0 -404 0 -5.507669 0.011416127 0.016564723201384055 0 -405 0 -5.95583 0.008152756 0.011810148833251818 0 -406 0 -4.1128182 0.0321975 0.047215431190392593 0 -407 0 -5.95583 0.008152756 0.011810148833251818 0 -408 0 -3.6910605 0.04380551 0.064624002308041828 0 -409 0 -4.71424 0.020646105 0.030097814327393193 0 -410 0 -5.95583 0.008152756 0.011810148833251818 0 +322 0 -4.697876 0.023884254 0.034875864219680933 0 +323 1 5.534587 0.9829864 0.024756652241414909 1 +324 0 -5.690696 0.011385534 0.016520077815512988 0 +325 0 -4.4049735 0.029654613 0.043429739587628399 0 +326 1 4.4437723 0.9618964 0.056046547132135686 1 +327 0 -6.1762524 0.007903442 0.011447553633965282 0 +328 1 3.1013708 0.90111274 0.15022048669228638 1 +329 1 7.778722 0.9968588 0.0045389615280207515 1 +330 1 5.7779264 0.98581505 0.020611091334643693 1 +331 0 -3.389574 0.061960425 0.092279305277741003 0 +332 0 -3.2830477 0.06682986 0.099787951160108662 0 +333 1 4.790513 0.9704533 0.043269272813164304 1 +334 1 5.7627764 0.98565334 0.020847762100166721 1 +335 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +336 1 5.4697657 0.98214364 0.025994057813121283 1 +337 0 -5.690696 0.011385534 0.016520077815512988 0 +338 0 -6.255251 0.007446876 0.010783773513399939 0 +339 1 5.63153 0.98417425 0.023014322412222076 1 +340 1 6.586154 0.9922691 0.011196669036113778 1 +341 0 -5.690696 0.011385534 0.016520077815512988 0 +342 0 -6.1653986 0.007968302 0.011541875502083741 0 +343 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +344 1 10.009007 0.9994206 0.00083616394161634133 1 +345 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +346 0 -3.4967813 0.057395987 0.085276271781417307 0 +347 0 -6.3551583 0.0069067827 0.0099989513904285149 0 +348 1 0.0005674362 0.46405905 1.1076196845341322 1 +349 1 3.947217 0.9454059 0.080994225997538957 1 +350 0 -4.0323086 0.03897188 0.057349447418246373 0 +351 0 -5.194286 0.016509674 0.024017233951091451 0 +352 0 0.16456318 0.49511734 0.98597995548356743 1 +353 1 8.388366 0.9980201 0.0028592052441175748 1 +354 0 -6.1762524 0.007903442 0.011447553633965282 0 +355 0 -4.4151993 0.029432077 0.043098914609063545 0 +356 1 -0.90889454 0.3027285 1.7239035738004502 0 +357 1 12.536469 0.9999149 0.00012280083891345824 1 +358 1 5.258912 0.9791091 0.03045845727908067 1 +359 1 5.3155527 0.97997063 0.029189577129909081 1 +360 1 14.975728 0.99998665 1.926218548588174E-05 1 +361 1 6.1216927 0.9890372 0.015903286835586925 1 +362 0 -3.7176933 0.048969198 0.072436027043799436 0 +363 0 -2.2717314 0.13368079 0.20702938763274298 0 +364 0 -5.194286 0.016509674 0.024017233951091451 0 +365 0 -5.679842 0.011478641 0.016655956359413658 0 +366 1 13.38573 0.9999553 6.4494935658442702E-05 1 +367 1 11.090333 0.9997449 0.00036808982615047284 1 +368 0 -6.1007133 0.00836595 0.012120283853048863 0 +369 0 -5.6941557 0.011356011 0.01647699535628969 0 +370 0 -4.1029515 0.037012495 0.054411016171145403 0 +371 0 -6.1007133 0.00836595 0.012120283853048863 0 +372 0 -4.4064746 0.029621843 0.04338101918277134 0 +373 0 -3.868682 0.04389939 0.064765655876225223 0 +374 0 -4.9028845 0.020512937 0.02990165675388249 0 +375 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +376 0 -6.1762524 0.007903442 0.011447553633965282 0 +377 0 -6.255251 0.007446876 0.010783773513399939 0 +378 0 -3.9458194 0.041506458 0.061159383455607043 0 +379 0 -2.4808912 0.11633939 0.17843572363943125 0 +380 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +381 1 10.022679 0.99942654 0.00082755984936018088 1 +382 0 -3.7903118 0.04646504 0.06864225989276064 0 +383 0 -6.1653986 0.007968302 0.011541875502083741 0 +384 0 -6.1653986 0.007968302 0.011541875502083741 0 +385 0 -3.9392 0.04170681 0.061460981007392071 0 +386 1 5.9550896 0.98757774 0.018033781674696592 1 +387 0 -2.518508 0.11343603 0.17370335971820122 0 +388 0 -5.4870176 0.013263867 0.019263755746596497 0 +389 0 -3.4613237 0.05886955 0.087533383405879489 0 +390 0 -5.8543224 0.010069123 0.014600303254977636 0 +391 1 9.892915 0.99936724 0.00091317285221328493 1 +392 0 -5.183432 0.016643979 0.024214259971593023 0 +393 0 -6.740807 0.0051630726 0.0074680348838617122 0 +394 0 -5.5028944 0.013107062 0.019034511373130159 0 +395 0 -5.183432 0.016643979 0.024214259971593023 0 +396 0 -4.697876 0.023884254 0.034875864219680933 0 +397 0 -5.172579 0.01677935 0.024412878962009357 0 +398 0 -4.8783455 0.02089054 0.03045793855632532 0 +399 0 -5.988451 0.009103315 0.013193450934389695 0 +400 1 9.690976 0.9992625 0.0010643631955075943 1 +401 0 -5.668988 0.0115725 0.016792944693470865 0 +402 0 -3.4049926 0.061283685 0.091238861570224067 0 +403 0 -4.296591 0.03211573 0.04709354230666267 0 +404 0 -5.712862 0.011197699 0.016245994747829652 0 +405 0 -6.1762524 0.007903442 0.011447553633965282 0 +406 0 -4.2942123 0.032171898 0.047177264146830024 0 +407 0 -6.1762524 0.007903442 0.011447553633965282 0 +408 0 -3.9098177 0.042607352 0.06281736772370089 0 +409 0 -4.9028845 0.020512937 0.02990165675388249 0 +410 0 -6.1762524 0.007903442 0.011447553633965282 0 411 0 ? ? ? 0 -412 1 9.230707 0.9987937 0.0017413429278913565 1 -413 0 -3.279101 0.05892835 0.087623528685202179 0 -414 1 6.7173805 0.9919376 0.011678759699519599 1 -415 0 -0.6668339 0.3123661 0.54028742034074773 0 -416 1 8.809383 0.9983402 0.0023965899325754819 1 -417 0 -5.95583 0.008152756 0.011810148833251818 0 -418 0 -1.8758616 0.15365383 0.24068022447923637 0 -419 0 -5.4421444 0.0119909365 0.017403818416010273 0 -420 0 -2.5893164 0.09557056 0.14492013592578398 0 -421 1 11.824856 0.99983126 0.00024346198385283565 1 -422 0 -2.8105893 0.08201394 0.12345585262262933 0 -423 0 -3.4726496 0.051294286 0.075967458128490192 0 -424 0 -5.4949865 0.011525216 0.016723931014560982 0 -425 1 14.817663 0.9999826 2.5109685538071405E-05 1 -426 0 -2.8241491 0.08124284 0.12224450219548814 0 -427 1 4.3530817 0.9534221 0.068812983796401356 1 -428 0 -5.95583 0.008152756 0.011810148833251818 0 -429 0 -5.4818344 0.011639432 0.016890641442649377 0 -430 0 -5.50395 0.011448008 0.01661124970571581 0 -431 0 -2.5834928 0.09595309 0.14553046360362973 0 -432 0 -3.8628197 0.03866103 0.056882875876087251 0 -433 0 -4.463106 0.024871154 0.036335236482012728 0 -434 0 5.008401 0.97114086 5.1148279570464927 1 -435 1 7.444332 0.9953392 0.0067398103254229808 1 -436 1 3.841199 0.9328069 0.10034962060198541 1 -437 0 -5.0209913 0.016430289 0.023900786364877163 0 -438 0 -3.822938 0.03980127 0.058595066713824417 0 -439 0 -4.5469956 0.02337358 0.034121287570165061 0 -440 1 10.154686 0.99940115 0.00086421363869066531 1 -441 0 -1.8604474 0.1551806 0.2432851360775424 0 -442 0 -4.9326286 0.017549375 0.025543191293877034 0 -443 0 -5.9932313 0.007926506 0.011481093714930413 0 -444 0 -2.442047 0.10567284 0.16112539871549494 0 -445 0 -5.9689827 0.008072471 0.011693375209389432 0 -446 0 -6.442978 0.0056482 0.0081717296030013944 0 -447 0 -4.5469956 0.02337358 0.034121287570165061 0 -448 0 -6.53364 0.0052747848 0.007630046968004383 0 -449 1 10.298004 0.99946284 0.00077516203531090419 1 -450 0 -4.01365 0.03462596 0.050840061269355086 0 -451 0 -4.5469956 0.02337358 0.034121287570165061 0 -452 0 -4.8841314 0.018195055 0.026491662451451816 0 -453 1 8.777969 0.99830025 0.0024543010625143486 1 -454 0 -5.6234684 0.010466129 0.015179003802371546 0 -455 1 0.81635284 0.58322746 0.77786945830087184 1 -456 1 10.487385 0.99953467 0.00067149052326743921 1 -457 1 9.062628 0.9986299 0.0019779518845872311 1 -458 0 -4.253397 0.029035252 0.042509176394131383 0 -459 0 -3.9597979 0.036017757 0.052921523163131159 0 -460 0 -3.93614 0.03664608 0.053862176491974861 0 -461 0 -3.6959996 0.043648843 0.064387644210923578 0 -462 0 -3.4621449 0.051683456 0.076559389501439229 0 -463 0 -4.823963 0.019028598 0.027717016988036743 0 -464 0 -5.0209913 0.016430289 0.023900786364877163 0 -465 1 9.7853985 0.99920774 0.0011434497150273235 1 -466 1 9.541931 0.99904716 0.0013753126588050381 1 -467 1 7.7145195 0.99619967 0.0054931654697981886 1 -468 0 -5.0209913 0.016430289 0.023900786364877163 0 -469 0 -5.6622314 0.010165937 0.014741404953300367 0 -470 0 -5.0985007 0.015506634 0.022546609667162744 0 -471 0 -3.4621449 0.051683456 0.076559389501439229 0 -472 0 -4.163662 0.031017102 0.04545689222198334 0 -473 0 -5.0209913 0.016430289 0.023900786364877163 0 -474 0 -4.5469956 0.02337358 0.034121287570165061 0 -475 0 -5.4949865 0.011525216 0.016723931014560982 0 -476 0 -4.7273927 0.020445328 0.029802079142657811 0 -477 0 -5.0209913 0.016430289 0.023900786364877163 0 -478 0 -4.4195695 0.025684817 0.037539546668792209 0 -479 1 8.32148 0.9975986 0.0034686710325429922 1 -480 0 -4.6376576 0.021854056 0.031878356794688936 0 -481 0 -3.0822616 0.06777515 0.10125012812840116 0 -482 1 15.481422 0.99998945 1.5220544890042414E-05 1 -483 1 10.906595 0.9996614 0.0004885994477486597 1 -484 0 -4.253397 0.029035252 0.042509176394131383 0 -485 0 -5.0232906 0.016402125 0.023859477325874637 0 -486 0 -5.0985007 0.015506634 0.022546609667162744 0 -487 1 13.475906 0.9999518 6.9568659593352019E-05 1 -488 1 1.3273249 0.67341036 0.57044218705778627 1 -489 1 -0.6232023 0.31951913 1.6460257744908038 0 -490 0 -6.442978 0.0056482 0.0081717296030013944 0 -491 1 6.6713343 0.9916534 0.012092158077135179 1 -492 0 -4.624505 0.022068355 0.032194467339038989 0 -493 1 9.948912 0.99930006 0.0010101496965434329 1 -494 0 0.9629116 0.60997653 1.3583671531547612 1 -495 0 -5.0985007 0.015506634 0.022546609667162744 0 -496 0 -6.53364 0.0052747848 0.007630046968004383 0 -497 0 -4.893565 0.018067656 0.02630447000187312 0 -498 0 -4.533843 0.02360242 0.034459376167277234 0 -499 0 -4.533843 0.02360242 0.034459376167277234 0 -500 0 -3.0987039 0.06699134 0.10003761775121778 0 -501 0 -4.533843 0.02360242 0.034459376167277234 0 -502 0 -4.2284155 0.029574301 0.043310338334976231 0 -503 0 -4.046695 0.033797733 0.049602857843819056 0 -504 0 -6.442978 0.0056482 0.0081717296030013944 0 -505 0 -5.2401342 0.013949007 0.020265837999788049 0 -506 1 10.447666 0.9995204 0.00069205216452113084 1 -507 0 -5.0937376 0.015561891 0.022627586922226627 0 -508 0 -4.5469956 0.02337358 0.034121287570165061 0 -509 0 -5.9689827 0.008072471 0.011693375209389432 0 -510 0 -6.442978 0.0056482 0.0081717296030013944 0 -511 0 -4.0598474 0.033473443 0.049118721706209283 0 -512 0 -4.5469956 0.02337358 0.034121287570165061 0 -513 0 -5.0985007 0.015506634 0.022546609667162744 0 -514 1 10.719856 0.9996099 0.00056292303371696174 1 -515 1 8.648027 0.9981245 0.0027083436166447333 1 -516 0 -6.53364 0.0052747848 0.007630046968004383 0 -517 0 -6.0464916 0.00761503 0.011028209313642047 0 -518 0 -4.8959603 0.018035451 0.026257153680743675 0 -519 1 6.535844 0.9907582 0.013395120416830476 1 -520 0 -6.3523164 0.0060478873 0.0087517485118454551 0 -521 0 -4.9303293 0.017579472 0.02558738782535638 0 -522 1 5.502533 0.97998166 0.029173343678465424 1 -523 1 7.699238 0.9961555 0.0055571295392400502 1 -524 0 -5.0078387 0.016592305 0.02413845106369562 0 -525 0 -5.189559 0.014486662 0.02105269812453239 0 -526 0 -5.0209913 0.016430289 0.023900786364877163 0 -527 0 -4.046695 0.033797733 0.049602857843819056 0 -528 0 -3.1803741 0.063221365 0.094219922377553209 0 -529 0 -4.624505 0.022068355 0.032194467339038989 0 -530 1 6.515936 0.9906189 0.013597970981112992 1 -531 0 -4.1128182 0.0321975 0.047215431190392593 0 -532 0 -5.559344 0.010982096 0.015931457121809788 0 -533 0 -5.0078387 0.016592305 0.02413845106369562 0 -534 0 -5.4818344 0.011639432 0.016890641442649377 0 -535 0 -4.5760565 0.022875605 0.033385855715514209 0 -536 0 -3.5726995 0.047724236 0.070548678886274141 0 -537 0 -3.279101 0.05892835 0.087623528685202179 0 -538 0 -4.533843 0.02360242 0.034459376167277234 0 -539 0 -3.5858517 0.047272857 0.069865003739744969 0 -540 0 -3.6101003 0.046451267 0.068621422453170058 0 -541 0 -5.4949865 0.011525216 0.016723931014560982 0 -542 0 -4.2921433 0.02821799 0.041295369834781166 0 -543 0 -4.533843 0.02360242 0.034459376167277234 0 -544 0 -4.589209 0.022653647 0.033058177993601175 0 -545 0 -4.0598474 0.033473443 0.049118721706209283 0 -546 1 11.390259 0.99976534 0.00033858757684800106 1 -547 0 -6.0596447 0.0075399973 0.010919133821803805 0 -548 0 -5.5856485 0.010767465 0.015618406057693391 0 -549 1 6.3187485 0.98912174 0.015780004831897315 1 -550 0 -5.0078387 0.016592305 0.02413845106369562 0 -551 0 -5.4686823 0.011754768 0.017059005370868869 0 -552 0 -3.1100621 0.0664548 0.099208215824527435 0 -553 0 -1.7353668 0.1680306 0.26539763634075109 0 -554 0 -5.4949865 0.011525216 0.016723931014560982 0 -555 0 -1.9254999 0.14882067 0.2324649753514042 0 -556 0 -3.4240565 0.053118054 0.078743527992808426 0 -557 0 -3.93614 0.03664608 0.053862176491974861 0 -558 0 -5.4818344 0.011639432 0.016890641442649377 0 -559 0 -4.0598474 0.033473443 0.049118721706209283 0 -560 0 -3.5726995 0.047724236 0.070548678886274141 0 -561 0 -3.5726995 0.047724236 0.070548678886274141 0 -562 0 -5.4686823 0.011754768 0.017059005370868869 0 -563 0 -5.0078387 0.016592305 0.02413845106369562 0 -564 0 -3.8496675 0.039033547 0.057442027050147004 0 -565 1 12.086092 0.9998616 0.00019968567641750592 1 -566 0 -4.2270923 0.029603124 0.043353188393335441 0 -567 0 -3.6343493 0.045643255 0.067399438854885543 0 -568 1 4.1473055 0.94597834 0.080120938914498979 1 -569 1 10.713882 0.9996081 0.00056550378256940248 1 -570 1 8.017664 0.99697804 0.0043663610015818432 1 -571 1 11.034658 0.99969274 0.00044335339890107668 1 -572 0 -5.0078387 0.016592305 0.02413845106369562 0 -573 0 -5.95583 0.008152756 0.011810148833251818 0 -574 1 5.950967 0.9856713 0.020821502205009505 1 -575 0 -3.279101 0.05892835 0.087623528685202179 0 -576 0 -4.0598474 0.033473443 0.049118721706209283 0 -577 0 -5.95583 0.008152756 0.011810148833251818 0 -578 0 -5.95583 0.008152756 0.011810148833251818 0 -579 0 -5.4686823 0.011754768 0.017059005370868869 0 -580 0 -3.7662487 0.041477654 0.061116029294295819 0 -581 1 8.417797 0.9977674 0.0032245780320862543 1 -582 1 7.90798 0.9967167 0.0047446254970359954 1 -583 0 -5.4949865 0.011525216 0.016723931014560982 0 -584 0 -2.9291954 0.075489886 0.11323899247123828 0 -585 0 -6.442978 0.0056482 0.0081717296030013944 0 -586 1 13.98102 0.9999671 4.7467992532759307E-05 1 -587 0 -3.8628197 0.03866103 0.056882875876087251 0 -588 1 5.463169 0.9793874 0.030048455526317758 1 -589 0 -4.5469956 0.02337358 0.034121287570165061 0 -590 1 3.9684038 0.9386083 0.0914048941554364 1 -591 1 5.966527 0.98583704 0.020578904318245012 1 -592 1 5.7801704 0.98372144 0.023678256090487241 1 -593 0 -4.253397 0.029035252 0.042509176394131383 0 -594 1 5.101776 0.97306114 0.039397632242971693 1 -595 0 -4.0598474 0.033473443 0.049118721706209283 0 -596 0 -4.2402444 0.02931786 0.042929146807739925 0 -597 0 -2.9855018 0.07256243 0.10867791847551245 0 -598 0 -5.0078387 0.016592305 0.02413845106369562 0 -599 0 -3.6294346 0.04580593 0.067645376058973333 0 -600 0 -5.0078387 0.016592305 0.02413845106369562 0 -601 0 -6.0464916 0.00761503 0.011028209313642047 0 -602 0 -4.533843 0.02360242 0.034459376167277234 0 -603 1 4.8058825 0.96650916 0.049144684334205956 1 -604 1 6.1928034 0.9880442 0.01735250950197232 1 -605 1 9.95545 0.9993035 0.0010051587149210726 1 -606 0 -4.715564 0.020625811 0.030067920071712086 0 -607 0 -6.442978 0.0056482 0.0081717296030013944 0 -608 1 11.148426 0.9997181 0.00040671030044872971 1 -609 0 -4.5469956 0.02337358 0.034121287570165061 0 -610 1 8.926191 0.99848074 0.0021934995672910755 1 -611 1 6.9109592 0.993031 0.010089302795775161 1 -612 1 16.893515 0.99999636 5.2454803872987944E-06 1 -613 0 -5.226982 0.014086902 0.02046760732470778 0 -614 0 -5.5724964 0.010874256 0.01577415696876637 0 -615 0 -3.9466453 0.036365785 0.05344247532455048 0 -616 0 -5.0078387 0.016592305 0.02413845106369562 0 +412 1 8.967401 0.9987234 0.0018429388177067811 1 +413 0 -3.424508 0.060437027 0.08993823578204467 0 +414 1 6.574872 0.9922032 0.011292519612784241 1 +415 0 -0.8179264 0.317499 0.5510969598656511 0 +416 1 8.593532 0.99830514 0.0024472377853153996 1 +417 0 -6.1762524 0.007903442 0.011447553633965282 0 +418 0 -2.0245438 0.15694046 0.24629357198382623 0 +419 0 -5.7038784 0.011273453 0.016356526232963797 0 +420 0 -2.7462745 0.09717632 0.14748383703212281 0 +421 1 11.613648 0.9998285 0.00024741825783100146 1 +422 0 -3.0301971 0.07984084 0.12004466735204 0 +423 0 -3.6295166 0.052182022 0.077318069673474266 0 +424 0 -5.668988 0.0115725 0.016792944693470865 0 +425 1 14.596736 0.9999822 2.571163541833834E-05 1 +426 0 -3.0489054 0.078803785 0.11841961170009874 0 +427 1 4.065505 0.949859 0.07421468857573911 1 +428 0 -6.1762524 0.007903442 0.011447553633965282 0 +429 0 -5.679842 0.011478641 0.016655956359413658 0 +430 0 -5.7589035 0.010817284 0.015691062821407739 0 +431 0 -2.7103744 0.09959342 0.15135150041447548 0 +432 0 -3.9977875 0.039965212 0.058841410857878557 0 +433 0 -4.6323185 0.025072282 0.036632835008259541 0 +434 0 4.7780943 0.9701818 5.0676641210003543 1 +435 1 7.255061 0.9953327 0.0067492273009859575 1 +436 1 3.4924908 0.9245983 0.11310142139105944 1 +437 0 -5.172579 0.01677935 0.024412878962009357 0 +438 0 -3.9676237 0.04085299 0.060176138516406297 0 +439 0 -4.6761684 0.024271425 0.035448215294650733 0 +440 1 10.246327 0.999516 0.00069841858984711258 1 +441 0 -1.9451714 0.16507754 0.2602858707972624 0 +442 0 -5.1030173 0.017673038 0.02572479741217518 0 +443 0 -6.204343 0.007737998 0.01120698718348863 0 +444 0 -2.501964 0.11470506 0.17576992479896 0 +445 0 -6.1653986 0.007968302 0.011541875502083741 0 +446 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +447 0 -4.6761684 0.024271425 0.035448215294650733 0 +448 0 -6.740807 0.0051630726 0.0074680348838617122 0 +449 1 10.062795 0.9994437 0.00080278035035997357 1 +450 0 -4.122161 0.03649629 0.053637871829137027 0 +451 0 -4.6761684 0.024271425 0.035448215294650733 0 +452 0 -5.0251284 0.018729242 0.027276825970618424 0 +453 1 8.612309 0.9983291 0.0024126109998507542 1 +454 0 -5.8618755 0.010012137 0.01451725632996664 0 +455 1 0.8015022 0.6139477 0.70381235666319208 1 +456 1 10.460095 0.9995885 0.00059380629791341746 1 +457 1 8.789279 0.99853885 0.0021095328973762531 1 +458 0 -4.395621 0.029859576 0.043734508612521997 0 +459 0 -4.115073 0.036685944 0.053921877565570465 0 +460 0 -4.0323086 0.03897188 0.057349447418246373 0 +461 0 -3.8391252 0.044850733 0.066201885567450408 0 +462 0 -3.5358982 0.05581057 0.082851763614760432 0 +463 0 -4.9797544 0.019372817 0.028223341371256296 0 +464 0 -5.172579 0.01677935 0.024412878962009357 0 +465 1 9.567504 0.9991901 0.0011689235541023761 1 +466 1 9.408446 0.99908626 0.0013188496529827891 1 +467 1 7.540141 0.99623746 0.0054384400286121778 1 +468 0 -5.172579 0.01677935 0.024412878962009357 0 +469 0 -5.8957047 0.009760794 0.014151024954576021 0 +470 0 -5.2624307 0.015690476 0.022816040551058049 0 +471 0 -3.5358982 0.05581057 0.082851763614760432 0 +472 0 -4.2587566 0.033020504 0.048442795745315334 0 +473 0 -5.172579 0.01677935 0.024412878962009357 0 +474 0 -4.6761684 0.024271425 0.035448215294650733 0 +475 0 -5.668988 0.0115725 0.016792944693470865 0 +476 0 -4.8920307 0.020679122 0.030146453152458324 0 +477 0 -5.172579 0.01677935 0.024412878962009357 0 +478 0 -4.563906 0.026373396 0.038559504719201797 0 +479 1 8.153737 0.99763507 0.0034159186232814122 1 +480 0 -4.755167 0.022891132 0.033408781065013729 0 +481 0 -3.205307 0.070605285 0.10563665461895037 0 +482 1 15.148742 0.99998826 1.6940390568188567E-05 1 +483 1 10.712245 0.99966013 0.00049040587839595544 1 +484 0 -4.395621 0.029859576 0.043734508612521997 0 +485 0 -5.182016 0.01666158 0.02424008433227327 0 +486 0 -5.2624307 0.015690476 0.022816040551058049 0 +487 1 13.332547 0.9999535 6.7074793056570611E-05 1 +488 1 1.1748514 0.678596 0.55937512683548662 1 +489 1 -0.7666216 0.32599702 1.6170692971874125 0 +490 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +491 1 6.429366 0.9913006 0.012605516096950381 1 +492 0 -4.766021 0.022707582 0.033137795160461427 0 +493 1 9.704236 0.99926984 0.0010537784951873425 1 +494 0 0.97516537 0.6446847 1.4928281696870613 1 +495 0 -5.2624307 0.015690476 0.022816040551058049 0 +496 0 -6.740807 0.0051630726 0.0074680348838617122 0 +497 0 -5.060316 0.018244626 0.026564504851636327 0 +498 0 -4.687022 0.02407708 0.035160889311831381 0 +499 0 -4.687022 0.02407708 0.035160889311831381 0 +500 0 -3.2086458 0.07043917 0.10537881324670834 0 +501 0 -4.687022 0.02407708 0.035160889311831381 0 +502 0 -4.3650923 0.030538203 0.044744047143781454 0 +503 0 -4.2014656 0.034437597 0.050558592955505986 0 +504 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +505 0 -5.4444637 0.013693328 0.019891801873291485 0 +506 1 10.169904 0.99948716 0.00074005914559640434 1 +507 0 -5.289411 0.015377302 0.022357096819201758 0 +508 0 -4.6761684 0.024271425 0.035448215294650733 0 +509 0 -6.1653986 0.007968302 0.011541875502083741 0 +510 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +511 0 -4.190612 0.034712594 0.050969538544379397 0 +512 0 -4.6761684 0.024271425 0.035448215294650733 0 +513 0 -5.2624307 0.015690476 0.022816040551058049 0 +514 1 10.608193 0.99963224 0.00053066406257628327 1 +515 1 8.427064 0.99807733 0.0027764921755513462 1 +516 0 -6.740807 0.0051630726 0.0074680348838617122 0 +517 0 -6.255251 0.007446876 0.010783773513399939 0 +518 0 -5.066511 0.018160589 0.026441017349862376 0 +519 1 6.2683277 0.9901806 0.014236394231357875 1 +520 0 -6.58281 0.0058171 0.0084168057396557267 0 +521 0 -5.09358 0.017797833 0.025908089833754346 0 +522 1 5.4384966 0.9817226 0.026612677067011484 1 +523 1 7.66887 0.9965865 0.0049330615131101723 1 +524 0 -5.183432 0.016643979 0.024214259971593023 0 +525 0 -5.3470583 0.014728642 0.021406976586218332 0 +526 0 -5.172579 0.01677935 0.024412878962009357 0 +527 0 -4.2014656 0.034437597 0.050558592955505986 0 +528 0 -3.296733 0.06618495 0.098791259074008708 0 +529 0 -4.766021 0.022707582 0.033137795160461427 0 +530 1 6.384506 0.99100196 0.013040178852573244 1 +531 0 -4.2942123 0.032171898 0.047177264146830024 0 +532 0 -5.7696943 0.010729991 0.01556375426175181 0 +533 0 -5.183432 0.016643979 0.024214259971593023 0 +534 0 -5.679842 0.011478641 0.016655956359413658 0 +535 0 -4.8076706 0.022016486 0.032117949435472748 0 +536 0 -3.7050557 0.049417876 0.073116822887907204 0 +537 0 -3.424508 0.060437027 0.08993823578204467 0 +538 0 -4.687022 0.02407708 0.035160889311831381 0 +539 0 -3.694202 0.049806327 0.073706493931953998 0 +540 0 -3.7331467 0.04842581 0.071611951298414134 0 +541 0 -5.668988 0.0115725 0.016792944693470865 0 +542 0 -4.451644 0.028652055 0.04193992145071157 0 +543 0 -4.687022 0.02407708 0.035160889311831381 0 +544 0 -4.796817 0.022194576 0.032380686011919603 0 +545 0 -4.190612 0.034712594 0.050969538544379397 0 +546 1 11.1768465 0.9997611 0.00034469440692551525 1 +547 0 -6.244397 0.007508018 0.010872647698333309 0 +548 0 -5.747987 0.01090631 0.015820910652078166 0 +549 1 6.1616106 0.98936087 0.015431255544328516 1 +550 0 -5.183432 0.016643979 0.024214259971593023 0 +551 0 -5.690696 0.011385534 0.016520077815512988 0 +552 0 -3.3047552 0.0658096 0.098211475293270387 0 +553 0 -1.9243431 0.16726808 0.26407597329557198 0 +554 0 -5.668988 0.0115725 0.016792944693470865 0 +555 0 -2.141615 0.1455381 0.22691194163080294 0 +556 0 -3.5548706 0.05505654 0.081700083179925997 0 +557 0 -4.0323086 0.03897188 0.057349447418246373 0 +558 0 -5.679842 0.011478641 0.016655956359413658 0 +559 0 -4.190612 0.034712594 0.050969538544379397 0 +560 0 -3.7050557 0.049417876 0.073116822887907204 0 +561 0 -3.7050557 0.049417876 0.073116822887907204 0 +562 0 -5.690696 0.011385534 0.016520077815512988 0 +563 0 -5.183432 0.016643979 0.024214259971593023 0 +564 0 -4.0086412 0.03965031 0.058368267792054167 0 +565 1 11.879315 0.9998598 0.00020226577557808291 1 +566 0 -4.417328 0.029385958 0.043030362779035429 0 +567 0 -3.772091 0.04708167 0.069575524341932163 0 +568 1 3.9970121 0.9473242 0.078069830447766844 1 +569 1 10.578958 0.999624 0.00054253528007564532 1 +570 1 7.813756 0.9969409 0.0044200969827829428 1 +571 1 10.822152 0.9996874 0.00045109501765938913 1 +572 0 -5.183432 0.016643979 0.024214259971593023 0 +573 0 -6.1762524 0.007903442 0.011447553633965282 0 +574 1 5.7168713 0.9851522 0.021581487544259318 1 +575 0 -3.424508 0.060437027 0.08993823578204467 0 +576 0 -4.190612 0.034712594 0.050969538544379397 0 +577 0 -6.1762524 0.007903442 0.011447553633965282 0 +578 0 -6.1762524 0.007903442 0.011447553633965282 0 +579 0 -5.690696 0.011385534 0.016520077815512988 0 +580 0 -3.9100647 0.042599704 0.062805842966436126 0 +581 1 8.170806 0.99766546 0.003371959755648659 1 +582 1 7.6789007 0.9966123 0.0048957002186025061 1 +583 0 -5.668988 0.0115725 0.016792944693470865 0 +584 0 -2.9875202 0.08225332 0.12383210773546874 0 +585 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +586 1 13.784331 0.999967 4.7639980852592516E-05 1 +587 0 -3.9977875 0.039965212 0.058841410857878557 0 +588 1 5.322174 0.98006904 0.029044711002758405 1 +589 0 -4.6761684 0.024271425 0.035448215294650733 0 +590 1 3.6173382 0.93094486 0.10323237563930997 1 +591 1 5.7292347 0.98528886 0.02138135152882949 1 +592 1 5.8126287 0.9861787 0.020079007413078769 1 +593 0 -4.395621 0.029859576 0.043734508612521997 0 +594 1 4.848382 0.97168714 0.041436222410741089 1 +595 0 -4.190612 0.034712594 0.050969538544379397 0 +596 0 -4.4064746 0.029621843 0.04338101918277134 0 +597 0 -3.14396 0.07372268 0.11048390991340731 0 +598 0 -5.183432 0.016643979 0.024214259971593023 0 +599 0 -3.801262 0.046098165 0.068087287518617035 0 +600 0 -5.183432 0.016643979 0.024214259971593023 0 +601 0 -6.255251 0.007446876 0.010783773513399939 0 +602 0 -4.687022 0.02407708 0.035160889311831381 0 +603 1 4.6091757 0.9662404 0.049545910540647636 1 +604 1 6.150976 0.9892756 0.015555637754274819 1 +605 1 9.757047 0.9992985 0.0010123870387376296 1 +606 0 -4.850649 0.021324908 0.031098112975602468 0 +607 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +608 1 10.875676 0.9996998 0.00043320333946758821 1 +609 0 -4.6761684 0.024271425 0.035448215294650733 0 +610 1 9.026143 0.998779 0.0017626085941359387 1 +611 1 6.6211424 0.9924702 0.010904303451904442 1 +612 1 16.623665 0.9999962 5.5034553246245386E-06 1 +613 0 -5.4553175 0.013582504 0.019729706483517399 0 +614 0 -5.7588406 0.010817795 0.015691808532505395 0 +615 0 -4.125927 0.0363959 0.053487562760631269 0 +616 0 -5.183432 0.016643979 0.024214259971593023 0 617 0 ? ? ? 0 -618 0 -4.533843 0.02360242 0.034459376167277234 0 -619 0 -4.0598474 0.033473443 0.049118721706209283 0 -620 0 -5.0078387 0.016592305 0.02413845106369562 0 -621 0 0.3560543 0.49671578 0.99055474235503638 1 -622 0 -2.2074018 0.12371333 0.19052518105806376 0 -623 0 -6.442978 0.0056482 0.0081717296030013944 0 -624 0 -3.8450818 0.03916424 0.057638251306253135 0 -625 0 -3.4678864 0.0514704 0.076235297856163745 0 -626 1 5.760166 0.98347664 0.024037311317372478 1 -627 0 -4.1699953 0.030873032 0.04524240583052884 0 -628 0 -5.9689827 0.008072471 0.011693375209389432 0 -629 0 -5.0209913 0.016430289 0.023900786364877163 0 -630 0 -3.358376 0.055680554 0.082653115934739244 0 -631 0 -4.0598474 0.033473443 0.049118721706209283 0 -632 0 -6.442978 0.0056482 0.0081717296030013944 0 -633 1 4.3299494 0.95263666 0.070002027657626642 1 -634 0 -5.4949865 0.011525216 0.016723931014560982 0 -635 0 -4.6141906 0.022237848 0.032444534178600062 0 -636 1 10.12752 0.9993887 0.00088219670683892097 1 -637 0 -2.4650183 0.1040373 0.1584894220773467 0 -638 0 -5.0209913 0.016430289 0.023900786364877163 0 -639 0 -3.93614 0.03664608 0.053862176491974861 0 -640 0 -4.4101357 0.025864512 0.037805650440469914 0 -641 0 -5.0078387 0.016592305 0.02413845106369562 0 -642 0 -5.0078387 0.016592305 0.02413845106369562 0 -643 0 -6.442978 0.0056482 0.0081717296030013944 0 -644 0 -5.9689827 0.008072471 0.011693375209389432 0 -645 0 -5.0078387 0.016592305 0.02413845106369562 0 -646 0 -6.021953 0.0077569974 0.011234611610460798 0 -647 0 -5.832123 0.0089476975 0.012966897604500274 0 -648 1 12.2362 0.9998765 0.00017818502951934642 1 -649 0 -5.0078387 0.016592305 0.02413845106369562 0 -650 0 -3.7496572 0.041980937 0.061873731573025101 0 -651 0 -5.2175484 0.014186638 0.02061355963523696 0 -652 0 -3.8628197 0.03866103 0.056882875876087251 0 -653 0 -4.533843 0.02360242 0.034459376167277234 0 -654 0 -4.520691 0.023833439 0.034800762077037174 0 -655 0 -5.0078387 0.016592305 0.02413845106369562 0 -656 0 -4.0598474 0.033473443 0.049118721706209283 0 -657 0 -0.4869156 0.3424043 0.6047272426002388 0 -658 1 9.086258 0.9986542 0.0019429058245893991 1 -659 0 -6.442978 0.0056482 0.0081717296030013944 0 -660 0 -5.95583 0.008152756 0.011810148833251818 0 -661 0 -4.046695 0.033797733 0.049602857843819056 0 -662 0 -5.3686323 0.012669892 0.018395573044372514 0 -663 0 -5.3686323 0.012669892 0.018395573044372514 0 -664 0 -4.3969836 0.026117079 0.038179751109026085 0 -665 0 -6.442978 0.0056482 0.0081717296030013944 0 -666 0 -3.496932 0.05040527 0.074616168514894637 0 -667 0 -4.520691 0.023833439 0.034800762077037174 0 -668 1 2.804366 0.8634336 0.21184286026666635 1 -669 1 8.147337 0.99726033 0.0039579294672795707 1 -670 1 6.4856215 0.99040276 0.013912762642680866 1 -671 0 -4.087837 0.03279327 0.048103809585535576 0 -672 0 -4.9946866 0.016755886 0.024378450658658433 0 -673 0 -3.9078827 0.037410382 0.055007231944439391 0 -674 0 -5.95583 0.008152756 0.011810148833251818 0 -675 0 -4.140195 0.03155662 0.046260392474474922 0 -676 0 -5.6622314 0.010165937 0.014741404953300367 0 -677 0 -4.5469956 0.02337358 0.034121287570165061 0 -678 0 -6.442978 0.0056482 0.0081717296030013944 0 -679 0 -5.9689827 0.008072471 0.011693375209389432 0 -680 1 16.78001 0.99999607 5.6754386418026423E-06 1 -681 1 9.801079 0.9992171 0.0011299384356403619 1 -682 0 -3.3756719 0.05499469 0.081605660780250541 0 -683 0 -6.442978 0.0056482 0.0081717296030013944 0 -684 0 -6.442978 0.0056482 0.0081717296030013944 0 -685 0 -6.442978 0.0056482 0.0081717296030013944 0 -686 0 -6.442978 0.0056482 0.0081717296030013944 0 -687 0 -4.613783 0.022244575 0.03245445849095939 0 -688 0 -5.0209913 0.016430289 0.023900786364877163 0 -689 0 -4.177704 0.030698553 0.044982689460987609 0 -690 0 -5.832123 0.0089476975 0.012966897604500274 0 -691 1 4.4967804 0.95803064 0.061856295892801602 1 -692 0 -5.4949865 0.011525216 0.016723931014560982 0 -693 0 -4.684228 0.021111501 0.03078355747963463 0 -694 0 -4.9545784 0.017264588 0.025125051561995987 0 -695 0 -5.9689827 0.008072471 0.011693375209389432 0 -696 1 6.7127876 0.9919097 0.011719331311161625 1 -697 1 5.064643 0.97231287 0.040507480679364792 1 -698 1 6.1036224 0.9872181 0.018559275661035816 1 +618 0 -4.687022 0.02407708 0.035160889311831381 0 +619 0 -4.190612 0.034712594 0.050969538544379397 0 +620 0 -5.183432 0.016643979 0.024214259971593023 0 +621 0 0.21396732 0.5044921 1.0130200478649019 1 +622 0 -2.2733846 0.13353552 0.20678748512685408 0 +623 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +624 0 -3.9476805 0.041450292 0.061074846290505025 0 +625 0 -3.656497 0.051178366 0.075791189781208843 0 +626 1 5.46276 0.9820502 0.02613135016883469 1 +627 0 -4.335535 0.03120946 0.045743316644172881 0 +628 0 -6.1653986 0.007968302 0.011541875502083741 0 +629 0 -5.172579 0.01677935 0.024412878962009357 0 +630 0 -3.506401 0.05700222 0.084673720830722193 0 +631 0 -4.190612 0.034712594 0.050969538544379397 0 +632 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +633 1 4.123639 0.95191944 0.071088615414448192 1 +634 0 -5.668988 0.0115725 0.016792944693470865 0 +635 0 -4.827346 0.021697205 0.031647030661752271 0 +636 1 9.884419 0.9993631 0.00091911002269828856 1 +637 0 -2.6317916 0.10507156 0.16015576751941168 0 +638 0 -5.172579 0.01677935 0.024412878962009357 0 +639 0 -4.0323086 0.03897188 0.057349447418246373 0 +640 0 -4.528719 0.027067967 0.039589069666512745 0 +641 0 -5.183432 0.016643979 0.024214259971593023 0 +642 0 -5.183432 0.016643979 0.024214259971593023 0 +643 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +644 0 -6.1653986 0.007968302 0.011541875502083741 0 +645 0 -5.183432 0.016643979 0.024214259971593023 0 +646 0 -6.268999 0.0073701376 0.010672237396626542 0 +647 0 -6.0179486 0.008903549 0.012902631287702191 0 +648 1 11.790706 0.9998501 0.00021628439499486233 1 +649 0 -5.183432 0.016643979 0.024214259971593023 0 +650 0 -3.8956623 0.043047808 0.063481243579512139 0 +651 0 -5.4201303 0.013945038 0.020260031880497513 0 +652 0 -3.9977875 0.039965212 0.058841410857878557 0 +653 0 -4.687022 0.02407708 0.035160889311831381 0 +654 0 -4.697876 0.023884254 0.034875864219680933 0 +655 0 -5.183432 0.016643979 0.024214259971593023 0 +656 0 -4.190612 0.034712594 0.050969538544379397 0 +657 0 -0.8079653 0.3191397 0.55456925788102518 0 +658 1 8.931214 0.99868786 0.0018942560714121602 1 +659 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +660 0 -6.1762524 0.007903442 0.011447553633965282 0 +661 0 -4.2014656 0.034437597 0.050558592955505986 0 +662 0 -5.615157 0.012049335 0.017489094794386224 0 +663 0 -5.615157 0.012049335 0.017489094794386224 0 +664 0 -4.5395722 0.026851853 0.039268645318208735 0 +665 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +666 0 -3.645947 0.051568627 0.07638470915813507 0 +667 0 -4.697876 0.023884254 0.034875864219680933 0 +668 1 2.434762 0.8460123 0.24124946607420433 1 +669 1 7.828244 0.9969743 0.0043717948862439853 1 +670 1 6.2092333 0.9897347 0.014886221130161663 1 +671 0 -4.2636833 0.032901306 0.048264967849262458 0 +672 0 -5.194286 0.016509674 0.024017233951091451 0 +673 0 -4.0920978 0.037307262 0.054852687996971815 0 +674 0 -6.1762524 0.007903442 0.011447553633965282 0 +675 0 -4.3309355 0.031315193 0.045900779860358822 0 +676 0 -5.8957047 0.009760794 0.014151024954576021 0 +677 0 -4.6761684 0.024271425 0.035448215294650733 0 +678 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +679 0 -6.1653986 0.007968302 0.011541875502083741 0 +680 1 16.591042 0.99999607 5.6754386418026423E-06 1 +681 1 9.56899 0.999191 0.0011676326392758554 1 +682 0 -3.5122313 0.056764804 0.084310542391200338 0 +683 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +684 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +685 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +686 0 -6.6618085 0.0054803924 0.0079282798352131763 0 +687 0 -4.7977834 0.022178661 0.032357205372034376 0 +688 0 -5.172579 0.01677935 0.024412878962009357 0 +689 0 -4.4534364 0.028614214 0.041883718439598289 0 +690 0 -6.0179486 0.008903549 0.012902631287702191 0 +691 1 4.169842 0.9534993 0.068696189508194894 1 +692 0 -5.668988 0.0115725 0.016792944693470865 0 +693 0 -4.8679504 0.021052541 0.030696664046168932 0 +694 0 -5.132524 0.017288381 0.025159981694104945 0 +695 0 -6.1653986 0.007968302 0.011541875502083741 0 +696 1 6.262699 0.99013895 0.01429709952126406 1 +697 1 4.767309 0.9699441 0.044026461927471366 1 +698 1 5.8011494 0.9860594 0.020253498150637261 1 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-out.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-out.txt index 5d98ed80f7..386483cdd6 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-out.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-out.txt @@ -8,11 +8,11 @@ Confusion table ||====================== PREDICTED || positive | negative | Recall TRUTH ||====================== - positive || 234 | 5 | 0.9791 + positive || 233 | 6 | 0.9749 negative || 12 | 432 | 0.9730 ||====================== -Precision || 0.9512 | 0.9886 | -OVERALL 0/1 ACCURACY: 0.975110 +Precision || 0.9510 | 0.9863 | +OVERALL 0/1 ACCURACY: 0.973646 LOG LOSS/instance: NaN Test-set entropy (prior Log-Loss/instance): 0.934003 LOG-LOSS REDUCTION (RIG): 0.000000 @@ -21,15 +21,15 @@ AUC: 0.996146 OVERALL RESULTS --------------------------------------- AUC: 0.996146 (0.0000) -Accuracy: 0.975110 (0.0000) -Positive precision: 0.951220 (0.0000) -Positive recall: 0.979079 (0.0000) -Negative precision: 0.988558 (0.0000) +Accuracy: 0.973646 (0.0000) +Positive precision: 0.951020 (0.0000) +Positive recall: 0.974895 (0.0000) +Negative precision: 0.986301 (0.0000) Negative recall: 0.972973 (0.0000) Log-loss: NaN (0.0000) Log-loss reduction: 0.000000 (0.0000) -F1 Score: 0.964948 (0.0000) -AUPRC: 0.992065 (0.0000) +F1 Score: 0.962810 (0.0000) +AUPRC: 0.992010 (0.0000) --------------------------------------- Warning: Data does not contain a probability column. Will not output the Log-loss column diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-rp.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-rp.txt index ac4bbeaf12..e3e77cb0a8 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-rp.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-rp.txt @@ -1,4 +1,4 @@ AveragedPerceptron AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings -0.996146 0.97511 0.95122 0.979079 0.988558 0.972973 NaN 0 0.964948 0.992065 AveragedPerceptron %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=AveragedPerceptron cali={} dout=%Output% data=%Data% out=%Output% seed=1 +0.996146 0.973646 0.95102 0.974895 0.986301 0.972973 NaN 0 0.96281 0.99201 AveragedPerceptron %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=AveragedPerceptron cali={} dout=%Output% data=%Data% out=%Output% seed=1 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration.txt b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration.txt index 3bf1ca0969..750c3c6a9d 100644 --- a/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration.txt +++ b/test/BaselineOutput/Common/AveragedPerceptron/netcoreapp/AveragedPerceptron-TrainTest-breast-cancer.nocalibration.txt @@ -1,700 +1,700 @@ Instance Label Score Assigned -0 0 -3.5726995 0 -1 0 3.6101465 1 -2 0 -4.070944 0 -3 0 2.470542 1 -4 0 -3.4358397 0 -5 1 12.382595 1 -6 0 -1.4209604 0 -7 0 -4.701088 0 -8 0 -4.6745405 0 -9 0 -4.406417 0 -10 0 -5.559344 0 -11 0 -5.4818344 0 -12 1 -0.14206886 0 -13 0 -4.5691886 0 -14 1 9.321611 1 -15 1 1.3856993 1 -16 0 -4.533843 0 -17 0 -4.046695 0 -18 1 7.8903713 1 -19 0 -3.0987039 0 -20 1 7.528511 1 -21 1 7.875206 1 -22 0 -5.0078387 0 +0 0 -3.7050557 0 +1 0 3.4661026 1 +2 0 -4.24041 0 +3 0 2.1493654 1 +4 0 -3.5576057 0 +5 1 12.155442 1 +6 0 -1.573504 0 +7 0 -4.9137383 0 +8 0 -4.903472 0 +9 0 -4.57476 0 +10 0 -5.7696943 0 +11 0 -5.679842 0 +12 1 -0.2966156 0 +13 0 -4.7757645 0 +14 1 9.220881 1 +15 1 1.1802778 1 +16 0 -4.687022 0 +17 0 -4.2014656 0 +18 1 7.81973 1 +19 0 -3.2086458 0 +20 1 7.5404253 1 +21 1 7.730527 1 +22 0 -5.183432 0 23 1 ? 0 -24 0 -5.4686823 0 -25 1 1.741828 1 -26 0 -4.9710746 0 -27 0 -4.0598474 0 -28 0 -5.4818344 0 -29 0 -5.8557806 0 -30 0 -5.0985007 0 -31 0 -4.9946866 0 -32 1 7.46414 1 -33 0 -4.689259 0 -34 0 -4.71424 0 -35 0 -5.4818344 0 -36 1 9.099108 1 -37 0 -1.113348 0 -38 1 6.140953 1 -39 1 2.5109034 1 +24 0 -5.690696 0 +25 1 1.6430082 1 +26 0 -5.1501684 0 +27 0 -4.190612 0 +28 0 -5.679842 0 +29 0 -6.1007133 0 +30 0 -5.2624307 0 +31 0 -5.194286 0 +32 1 7.2332277 1 +33 0 -4.872356 0 +34 0 -4.9028845 0 +35 0 -5.679842 0 +36 1 8.852162 1 +37 0 -1.2331572 0 +38 1 6.101239 1 +39 1 2.2340298 1 40 0 ? 0 -41 1 3.3300762 1 -42 1 8.577511 1 -43 1 0.49126053 1 -44 1 8.255751 1 -45 0 -5.6322193 0 -46 1 4.5673857 1 -47 0 -5.95583 0 -48 0 -3.4358397 0 -49 1 5.3666544 1 -50 1 2.5949678 1 -51 1 0.12595749 1 -52 1 5.2992115 1 -53 1 8.407228 1 -54 1 7.649311 1 -55 1 4.478709 1 -56 1 5.5541325 1 -57 1 1.6657066 1 -58 1 2.5265894 1 -59 1 1.7368536 1 -60 1 2.3288136 1 -61 0 -5.5060835 0 -62 1 6.380088 1 -63 1 0.3348999 1 -64 0 -5.95583 0 -65 1 3.8072634 1 -66 0 -4.046695 0 -67 1 4.218014 1 -68 1 10.826723 1 -69 0 -5.2716546 0 -70 0 -3.4726496 0 -71 1 7.895048 1 -72 0 -2.1755848 0 -73 1 8.9055195 1 -74 1 2.5993576 1 -75 0 -4.0411606 0 -76 0 -5.075033 0 -77 0 -3.4995675 0 -78 0 -3.6211967 0 -79 0 -5.3911724 0 -80 0 -2.7157316 0 -81 0 -4.2284155 0 -82 0 -3.4452734 0 -83 0 -2.1223516 0 -84 1 9.694054 1 -85 1 6.2895603 1 -86 1 2.6168842 1 -87 1 6.91914 1 -88 0 -4.046695 0 -89 0 -5.085745 0 -90 0 -5.4686823 0 -91 0 -5.189559 0 -92 0 -4.046695 0 -93 0 -5.95583 0 -94 0 -4.9946866 0 -95 0 -5.4686823 0 -96 0 -5.663555 0 -97 0 -3.5726995 0 -98 1 8.590233 1 -99 1 10.917194 1 -100 1 4.8476696 1 -101 1 -0.842803 0 -102 0 -3.7530966 0 -103 1 1.7746449 1 -104 1 12.140858 1 -105 1 2.5560703 1 -106 1 9.259369 1 -107 1 6.720646 1 -108 0 -5.5617743 0 -109 1 6.871727 1 -110 0 -2.766693 0 -111 1 3.848031 1 -112 1 9.42577 1 -113 1 9.506622 1 -114 0 -3.0727453 0 -115 0 -4.643991 0 -116 0 -0.6618881 0 -117 1 9.617277 1 -118 0 -5.3621607 0 -119 0 -3.9435177 0 -120 0 -4.8696556 0 -121 0 -3.469522 0 -122 1 9.680523 1 -123 1 3.8165932 1 -124 1 7.6522446 1 -125 0 -5.95583 0 -126 1 8.564951 1 -127 0 -4.520691 0 -128 1 4.848981 1 -129 0 -5.717684 0 -130 0 -3.4726496 0 -131 0 -4.9946866 0 -132 1 8.60223 1 -133 0 -4.810811 0 -134 0 -4.917177 0 -135 0 -2.7288966 0 -136 0 -4.533843 0 -137 0 -5.4949865 0 -138 0 -4.2402444 0 +41 1 3.2122545 1 +42 1 8.198594 1 +43 1 0.22008419 1 +44 1 7.915265 1 +45 0 -5.8607707 0 +46 1 4.2778234 1 +47 0 -6.1762524 0 +48 0 -3.5576057 0 +49 1 5.1043253 1 +50 1 2.3847685 1 +51 1 -0.002776146 0 +52 1 5.171505 1 +53 1 8.174706 1 +54 1 7.5763617 1 +55 1 4.3368845 1 +56 1 5.228385 1 +57 1 1.4711056 1 +58 1 2.3726006 1 +59 1 1.6023321 1 +60 1 2.1442842 1 +61 0 -5.7187867 0 +62 1 6.0357714 1 +63 1 0.09457874 1 +64 0 -6.1762524 0 +65 1 3.6413336 1 +66 0 -4.2014656 0 +67 1 3.980445 1 +68 1 10.612863 1 +69 0 -5.4978714 0 +70 0 -3.6295166 0 +71 1 7.5565577 1 +72 0 -2.4480972 0 +73 1 8.905442 1 +74 1 2.4551315 1 +75 0 -4.274302 0 +76 0 -5.334609 0 +77 0 -3.644158 0 +78 0 -3.7829442 0 +79 0 -5.6008434 0 +80 0 -2.9382496 0 +81 0 -4.3650923 0 +82 0 -3.5927935 0 +83 0 -2.2981849 0 +84 1 9.3922205 1 +85 1 5.7727385 1 +86 1 2.3984623 1 +87 1 6.6674128 1 +88 0 -4.2014656 0 +89 0 -5.2789135 0 +90 0 -5.690696 0 +91 0 -5.3470583 0 +92 0 -4.2014656 0 +93 0 -6.1762524 0 +94 0 -5.194286 0 +95 0 -5.690696 0 +96 0 -5.8434687 0 +97 0 -3.7050557 0 +98 1 8.223899 1 +99 1 10.822373 1 +100 1 4.636462 1 +101 1 -1.0467243 0 +102 0 -3.9209185 0 +103 1 1.630683 1 +104 1 11.827565 1 +105 1 2.3993664 1 +106 1 9.16854 1 +107 1 6.4510603 1 +108 0 -5.7906036 0 +109 1 6.7095194 1 +110 0 -3.014533 0 +111 1 3.7277098 1 +112 1 9.536014 1 +113 1 9.218173 1 +114 0 -3.277872 0 +115 0 -4.8319454 0 +116 0 -0.7799158 0 +117 1 9.464227 1 +118 0 -5.5888796 0 +119 0 -4.1281476 0 +120 0 -5.088218 0 +121 0 -3.6317377 0 +122 1 9.371118 1 +123 1 3.5896034 1 +124 1 7.4270334 1 +125 0 -6.1762524 0 +126 1 8.480126 1 +127 0 -4.697876 0 +128 1 4.7593575 1 +129 0 -6.02982 0 +130 0 -3.6295166 0 +131 0 -5.194286 0 +132 1 8.352777 1 +133 0 -4.9906077 0 +134 0 -5.1044335 0 +135 0 -2.8619413 0 +136 0 -4.687022 0 +137 0 -5.668988 0 +138 0 -4.4064746 0 139 0 ? 0 -140 0 -5.4949865 0 -141 0 -5.9689827 0 -142 1 4.4324036 1 -143 0 -4.643991 0 -144 0 -5.4818344 0 +140 0 -5.668988 0 +141 0 -6.1653986 0 +142 1 4.307641 1 +143 0 -4.8319454 0 +144 0 -5.679842 0 145 0 ? 0 -146 1 1.3394356 1 -147 0 -5.4154215 0 -148 0 -1.012373 0 -149 1 11.461615 1 -150 0 -5.559344 0 -151 1 5.006485 1 -152 1 9.715748 1 -153 0 -4.1214976 0 -154 0 -6.442978 0 -155 1 3.7769232 1 -156 0 -5.5348053 0 -157 0 -4.9946866 0 +146 1 1.0868425 1 +147 0 -5.639788 0 +148 0 -1.259141 0 +149 1 11.153153 1 +150 0 -5.7696943 0 +151 1 4.855405 1 +152 1 9.511897 1 +153 0 -4.257647 0 +154 0 -6.6618085 0 +155 1 3.6720238 1 +156 0 -5.7834425 0 +157 0 -5.194286 0 158 0 ? 0 -159 1 12.346203 1 -160 1 9.039494 1 -161 0 -3.8496675 0 -162 0 -4.520691 0 -163 0 -3.387055 0 +159 1 12.03571 1 +160 1 8.89737 1 +161 0 -4.0086412 0 +162 0 -4.697876 0 +163 0 -3.6110554 0 164 0 ? 0 -165 0 -3.3999205 0 -166 1 7.976183 1 -167 1 8.355644 1 -168 0 -4.520691 0 -169 0 -6.2282124 0 -170 0 -5.4949865 0 -171 0 -5.4686823 0 -172 0 -5.95583 0 -173 1 15.1560135 1 -174 1 6.1769247 1 -175 1 7.842922 1 -176 0 -4.9946866 0 -177 1 4.766121 1 -178 0 -4.046695 0 -179 1 2.290575 1 -180 0 -5.559344 0 -181 0 -6.442978 0 -182 0 -3.0987039 0 -183 1 9.159964 1 -184 1 6.2014647 1 -185 0 -5.0853486 0 -186 1 5.7654104 1 -187 1 13.977451 1 -188 1 9.065283 1 -189 0 -4.7540584 0 -190 1 11.957218 1 -191 1 10.956873 1 -192 0 -4.0598474 0 -193 0 -5.4686823 0 -194 0 -4.520691 0 -195 0 -4.046695 0 -196 0 6.8652763 1 -197 0 -2.6564164 0 -198 0 -6.442978 0 -199 0 -5.0078387 0 -200 1 10.36586 1 -201 1 9.869495 1 -202 0 -5.4686823 0 -203 0 -3.5726995 0 -204 0 -5.4686823 0 -205 1 12.086601 1 -206 1 5.944169 1 -207 0 -5.559344 0 -208 0 -5.559344 0 -209 0 -3.6633615 0 -210 1 14.534113 1 -211 1 9.64962 1 -212 0 -5.4686823 0 -213 1 14.529058 1 -214 1 13.868914 1 -215 1 7.643734 1 -216 0 -5.95583 0 -217 0 -5.4686823 0 -218 1 7.8867817 1 -219 0 -2.511506 0 -220 0 -5.1632547 0 -221 1 10.395218 1 -222 1 -2.214662 0 -223 1 5.7424126 1 -224 1 9.995327 1 -225 0 -5.95583 0 -226 1 10.225868 1 -227 1 7.459608 1 -228 0 -5.559344 0 -229 1 12.666515 1 -230 1 6.1583214 1 -231 1 8.623034 1 -232 0 1.2822819 1 -233 1 6.3825197 1 -234 0 -2.8964381 0 +165 0 -3.5511756 0 +166 1 7.750617 1 +167 1 8.140458 1 +168 0 -4.697876 0 +169 0 -6.418877 0 +170 0 -5.668988 0 +171 0 -5.690696 0 +172 0 -6.1762524 0 +173 1 14.842399 1 +174 1 6.0863256 1 +175 1 7.5728846 1 +176 0 -5.194286 0 +177 1 4.2855787 1 +178 0 -4.2014656 0 +179 1 2.15761 1 +180 0 -5.7696943 0 +181 0 -6.6618085 0 +182 0 -3.2086458 0 +183 1 8.944306 1 +184 1 6.0725775 1 +185 0 -5.2732844 0 +186 1 5.448736 1 +187 1 13.897843 1 +188 1 8.907805 1 +189 0 -5.0173383 0 +190 1 11.761691 1 +191 1 10.76023 1 +192 0 -4.190612 0 +193 0 -5.690696 0 +194 0 -4.697876 0 +195 0 -4.2014656 0 +196 0 6.7605906 1 +197 0 -2.7392664 0 +198 0 -6.6618085 0 +199 0 -5.183432 0 +200 1 10.20231 1 +201 1 9.669415 1 +202 0 -5.690696 0 +203 0 -3.7050557 0 +204 0 -5.690696 0 +205 1 11.778428 1 +206 1 5.671749 1 +207 0 -5.7696943 0 +208 0 -5.7696943 0 +209 0 -3.7840543 0 +210 1 14.311627 1 +211 1 9.471307 1 +212 0 -5.690696 0 +213 1 14.345925 1 +214 1 13.768859 1 +215 1 7.442033 1 +216 0 -6.1762524 0 +217 0 -5.690696 0 +218 1 7.5836535 1 +219 0 -2.6475506 0 +220 0 -5.368766 0 +221 1 10.2217455 1 +222 1 -2.2922978 0 +223 1 5.585368 1 +224 1 9.954375 1 +225 0 -6.1762524 0 +226 1 10.120683 1 +227 1 7.2224054 1 +228 0 -5.7696943 0 +229 1 12.354081 1 +230 1 6.003766 1 +231 1 8.352307 1 +232 0 1.1177626 1 +233 1 6.363435 1 +234 0 -3.1496835 0 235 0 ? 0 -236 1 11.420414 1 -237 1 6.535795 1 -238 1 12.422876 1 -239 1 5.9025297 1 -240 0 -2.0179915 0 -241 0 -4.0004973 0 -242 0 -4.9946866 0 -243 0 -2.6953988 0 -244 0 -5.4686823 0 -245 0 -2.817525 0 -246 1 11.424002 1 -247 1 3.104393 1 -248 0 -3.0615559 0 +236 1 11.114438 1 +237 1 6.258853 1 +238 1 12.123448 1 +239 1 5.86524 1 +240 0 -2.1425428 0 +241 0 -4.1330147 0 +242 0 -5.194286 0 +243 0 -2.8897204 0 +244 0 -5.690696 0 +245 0 -2.9256601 0 +246 1 11.219155 1 +247 1 3.0079794 1 +248 0 -3.2373443 0 249 0 ? 0 -250 0 -6.021953 0 -251 1 8.872498 1 -252 0 4.5387735 1 -253 1 8.577511 1 -254 1 6.380088 1 -255 1 4.052039 1 -256 0 -5.4949865 0 -257 0 -5.0078387 0 -258 0 -4.520691 0 -259 0 2.9647484 1 -260 1 9.870926 1 -261 1 12.206299 1 -262 1 9.653839 1 -263 1 8.981979 1 -264 1 5.664708 1 -265 0 -2.4948754 0 -266 1 7.3661633 1 -267 1 3.3009605 1 -268 1 9.372967 1 -269 0 -5.4686823 0 -270 1 6.031377 1 -271 0 -3.5726995 0 -272 1 3.3009605 1 -273 1 0.21747208 1 -274 0 -4.323663 0 +250 0 -6.268999 0 +251 1 8.957709 1 +252 0 4.4631453 1 +253 1 8.198594 1 +254 1 6.0357714 1 +255 1 3.8202343 1 +256 0 -5.668988 0 +257 0 -5.183432 0 +258 0 -4.697876 0 +259 0 2.6695328 1 +260 1 9.885512 1 +261 1 11.814006 1 +262 1 9.426362 1 +263 1 9.02839 1 +264 1 5.4595184 1 +265 0 -2.6705885 0 +266 1 7.2168283 1 +267 1 3.1091843 1 +268 1 9.093087 1 +269 0 -5.690696 0 +270 1 5.858514 1 +271 0 -3.7050557 0 +272 1 3.1091843 1 +273 1 0.07871342 1 +274 0 -4.5050516 0 275 0 ? 0 -276 0 -5.0078387 0 -277 0 -5.95583 0 -278 0 -5.4686823 0 -279 1 7.127905 1 -280 0 -4.520691 0 -281 0 -4.689259 0 -282 1 4.4381237 1 -283 1 6.0636253 1 -284 1 7.431343 1 -285 1 14.218481 1 -286 1 15.281263 1 -287 0 -4.917177 0 -288 1 2.2163515 1 -289 1 8.312019 1 -290 0 -6.442978 0 -291 0 -5.4686823 0 +276 0 -5.183432 0 +277 0 -6.1762524 0 +278 0 -5.690696 0 +279 1 6.9710016 1 +280 0 -4.697876 0 +281 0 -4.872356 0 +282 1 4.3149786 1 +283 1 5.990473 1 +284 1 7.4265547 1 +285 1 13.994821 1 +286 1 15.019143 1 +287 0 -5.1044335 0 +288 1 2.0468407 1 +289 1 8.113109 1 +290 0 -6.6618085 0 +291 0 -5.690696 0 292 1 ? 0 -293 1 5.542122 1 +293 1 5.5823574 1 294 0 ? 0 -295 1 7.7866364 1 -296 0 1.823431 1 +295 1 7.634673 1 +296 0 1.6353331 1 297 0 ? 0 -298 0 -2.725597 0 -299 1 7.8274364 1 -300 1 7.348074 1 -301 0 -5.4686823 0 -302 1 15.735764 1 -303 0 -5.4686823 0 -304 1 5.9607973 1 -305 1 8.459469 1 -306 0 -5.4686823 0 -307 0 -5.4686823 0 -308 1 7.4225903 1 -309 0 -1.7474074 0 -310 0 -5.3911724 0 -311 0 -6.442978 0 -312 1 3.6294699 1 -313 0 -6.442978 0 -314 0 -6.0464916 0 +298 0 -2.8376803 0 +299 1 7.8606234 1 +300 1 7.1416483 1 +301 0 -5.690696 0 +302 1 15.431024 1 +303 0 -5.690696 0 +304 1 6.007621 1 +305 1 8.327315 1 +306 0 -5.690696 0 +307 0 -5.690696 0 +308 1 7.1634607 1 +309 0 -1.8969002 0 +310 0 -5.6008434 0 +311 0 -6.6618085 0 +312 1 3.4877834 1 +313 0 -6.6618085 0 +314 0 -6.255251 0 315 0 ? 0 -316 1 3.6177397 1 -317 1 9.215706 1 -318 0 -5.1966968 0 -319 0 2.6369457 1 -320 1 7.3824844 1 +316 1 3.4135065 1 +317 1 9.002616 1 +318 0 -5.4537 0 +319 0 2.4164848 1 +320 1 7.184394 1 321 0 ? 0 -322 0 -4.520691 0 -323 1 5.5612926 1 -324 0 -5.4686823 0 -325 0 -4.1927576 0 -326 1 4.4103804 1 -327 0 -5.95583 0 -328 1 3.373887 1 -329 1 7.8321342 1 -330 1 5.8562517 1 -331 0 -3.2490892 0 -332 0 -3.1363668 0 -333 1 4.914962 1 -334 1 5.9119453 1 -335 0 -6.442978 0 -336 1 5.54352 1 -337 0 -5.4686823 0 -338 0 -6.0464916 0 -339 1 5.684024 1 -340 1 6.620782 1 -341 0 -5.4686823 0 -342 0 -5.9689827 0 -343 0 -6.442978 0 -344 1 10.162451 1 -345 0 -6.442978 0 -346 0 -3.335825 0 -347 0 -6.139584 0 -348 1 0.15727425 1 -349 1 4.0622606 1 -350 0 -3.93614 0 -351 0 -4.9946866 0 -352 0 0.4719286 1 -353 1 8.696344 1 -354 0 -5.95583 0 -355 0 -4.246154 0 -356 1 -0.69921684 0 -357 1 12.852016 1 -358 1 5.5822067 1 -359 1 5.3672857 1 -360 1 15.333874 1 -361 1 6.31769 1 -362 0 -3.5059962 0 -363 0 -2.065846 0 -364 0 -4.9946866 0 -365 0 -5.4818344 0 -366 1 13.694569 1 -367 1 11.299244 1 -368 0 -5.8557806 0 -369 0 -5.4592943 0 -370 0 -3.8947306 0 -371 0 -5.8557806 0 -372 0 -4.2402444 0 -373 0 -3.7544198 0 -374 0 -4.71424 0 -375 0 -6.442978 0 -376 0 -5.95583 0 -377 0 -6.0464916 0 -378 0 -3.8038664 0 -379 0 -2.2557268 0 -380 0 -6.442978 0 -381 1 10.076408 1 -382 0 -3.5972104 0 -383 0 -5.9689827 0 -384 0 -5.9689827 0 -385 0 -3.7061968 0 -386 1 6.0875874 1 -387 0 -2.33456 0 -388 0 -5.2848067 0 -389 0 -3.322411 0 -390 0 -5.6504025 0 -391 1 10.030338 1 -392 0 -5.0078387 0 -393 0 -6.53364 0 -394 0 -5.241206 0 -395 0 -5.0078387 0 -396 0 -4.520691 0 -397 0 -5.0209913 0 -398 0 -4.683385 0 -399 0 -5.7283545 0 -400 1 10.056744 1 -401 0 -5.4949865 0 -402 0 -3.2177973 0 -403 0 -4.145746 0 -404 0 -5.507669 0 -405 0 -5.95583 0 -406 0 -4.1128182 0 -407 0 -5.95583 0 -408 0 -3.6910605 0 -409 0 -4.71424 0 -410 0 -5.95583 0 +322 0 -4.697876 0 +323 1 5.534587 1 +324 0 -5.690696 0 +325 0 -4.4049735 0 +326 1 4.4437723 1 +327 0 -6.1762524 0 +328 1 3.1013708 1 +329 1 7.778722 1 +330 1 5.7779264 1 +331 0 -3.389574 0 +332 0 -3.2830477 0 +333 1 4.790513 1 +334 1 5.7627764 1 +335 0 -6.6618085 0 +336 1 5.4697657 1 +337 0 -5.690696 0 +338 0 -6.255251 0 +339 1 5.63153 1 +340 1 6.586154 1 +341 0 -5.690696 0 +342 0 -6.1653986 0 +343 0 -6.6618085 0 +344 1 10.009007 1 +345 0 -6.6618085 0 +346 0 -3.4967813 0 +347 0 -6.3551583 0 +348 1 0.0005674362 1 +349 1 3.947217 1 +350 0 -4.0323086 0 +351 0 -5.194286 0 +352 0 0.16456318 1 +353 1 8.388366 1 +354 0 -6.1762524 0 +355 0 -4.4151993 0 +356 1 -0.90889454 0 +357 1 12.536469 1 +358 1 5.258912 1 +359 1 5.3155527 1 +360 1 14.975728 1 +361 1 6.1216927 1 +362 0 -3.7176933 0 +363 0 -2.2717314 0 +364 0 -5.194286 0 +365 0 -5.679842 0 +366 1 13.38573 1 +367 1 11.090333 1 +368 0 -6.1007133 0 +369 0 -5.6941557 0 +370 0 -4.1029515 0 +371 0 -6.1007133 0 +372 0 -4.4064746 0 +373 0 -3.868682 0 +374 0 -4.9028845 0 +375 0 -6.6618085 0 +376 0 -6.1762524 0 +377 0 -6.255251 0 +378 0 -3.9458194 0 +379 0 -2.4808912 0 +380 0 -6.6618085 0 +381 1 10.022679 1 +382 0 -3.7903118 0 +383 0 -6.1653986 0 +384 0 -6.1653986 0 +385 0 -3.9392 0 +386 1 5.9550896 1 +387 0 -2.518508 0 +388 0 -5.4870176 0 +389 0 -3.4613237 0 +390 0 -5.8543224 0 +391 1 9.892915 1 +392 0 -5.183432 0 +393 0 -6.740807 0 +394 0 -5.5028944 0 +395 0 -5.183432 0 +396 0 -4.697876 0 +397 0 -5.172579 0 +398 0 -4.8783455 0 +399 0 -5.988451 0 +400 1 9.690976 1 +401 0 -5.668988 0 +402 0 -3.4049926 0 +403 0 -4.296591 0 +404 0 -5.712862 0 +405 0 -6.1762524 0 +406 0 -4.2942123 0 +407 0 -6.1762524 0 +408 0 -3.9098177 0 +409 0 -4.9028845 0 +410 0 -6.1762524 0 411 0 ? 0 -412 1 9.230707 1 -413 0 -3.279101 0 -414 1 6.7173805 1 -415 0 -0.6668339 0 -416 1 8.809383 1 -417 0 -5.95583 0 -418 0 -1.8758616 0 -419 0 -5.4421444 0 -420 0 -2.5893164 0 -421 1 11.824856 1 -422 0 -2.8105893 0 -423 0 -3.4726496 0 -424 0 -5.4949865 0 -425 1 14.817663 1 -426 0 -2.8241491 0 -427 1 4.3530817 1 -428 0 -5.95583 0 -429 0 -5.4818344 0 -430 0 -5.50395 0 -431 0 -2.5834928 0 -432 0 -3.8628197 0 -433 0 -4.463106 0 -434 0 5.008401 1 -435 1 7.444332 1 -436 1 3.841199 1 -437 0 -5.0209913 0 -438 0 -3.822938 0 -439 0 -4.5469956 0 -440 1 10.154686 1 -441 0 -1.8604474 0 -442 0 -4.9326286 0 -443 0 -5.9932313 0 -444 0 -2.442047 0 -445 0 -5.9689827 0 -446 0 -6.442978 0 -447 0 -4.5469956 0 -448 0 -6.53364 0 -449 1 10.298004 1 -450 0 -4.01365 0 -451 0 -4.5469956 0 -452 0 -4.8841314 0 -453 1 8.777969 1 -454 0 -5.6234684 0 -455 1 0.81635284 1 -456 1 10.487385 1 -457 1 9.062628 1 -458 0 -4.253397 0 -459 0 -3.9597979 0 -460 0 -3.93614 0 -461 0 -3.6959996 0 -462 0 -3.4621449 0 -463 0 -4.823963 0 -464 0 -5.0209913 0 -465 1 9.7853985 1 -466 1 9.541931 1 -467 1 7.7145195 1 -468 0 -5.0209913 0 -469 0 -5.6622314 0 -470 0 -5.0985007 0 -471 0 -3.4621449 0 -472 0 -4.163662 0 -473 0 -5.0209913 0 -474 0 -4.5469956 0 -475 0 -5.4949865 0 -476 0 -4.7273927 0 -477 0 -5.0209913 0 -478 0 -4.4195695 0 -479 1 8.32148 1 -480 0 -4.6376576 0 -481 0 -3.0822616 0 -482 1 15.481422 1 -483 1 10.906595 1 -484 0 -4.253397 0 -485 0 -5.0232906 0 -486 0 -5.0985007 0 -487 1 13.475906 1 -488 1 1.3273249 1 -489 1 -0.6232023 0 -490 0 -6.442978 0 -491 1 6.6713343 1 -492 0 -4.624505 0 -493 1 9.948912 1 -494 0 0.9629116 1 -495 0 -5.0985007 0 -496 0 -6.53364 0 -497 0 -4.893565 0 -498 0 -4.533843 0 -499 0 -4.533843 0 -500 0 -3.0987039 0 -501 0 -4.533843 0 -502 0 -4.2284155 0 -503 0 -4.046695 0 -504 0 -6.442978 0 -505 0 -5.2401342 0 -506 1 10.447666 1 -507 0 -5.0937376 0 -508 0 -4.5469956 0 -509 0 -5.9689827 0 -510 0 -6.442978 0 -511 0 -4.0598474 0 -512 0 -4.5469956 0 -513 0 -5.0985007 0 -514 1 10.719856 1 -515 1 8.648027 1 -516 0 -6.53364 0 -517 0 -6.0464916 0 -518 0 -4.8959603 0 -519 1 6.535844 1 -520 0 -6.3523164 0 -521 0 -4.9303293 0 -522 1 5.502533 1 -523 1 7.699238 1 -524 0 -5.0078387 0 -525 0 -5.189559 0 -526 0 -5.0209913 0 -527 0 -4.046695 0 -528 0 -3.1803741 0 -529 0 -4.624505 0 -530 1 6.515936 1 -531 0 -4.1128182 0 -532 0 -5.559344 0 -533 0 -5.0078387 0 -534 0 -5.4818344 0 -535 0 -4.5760565 0 -536 0 -3.5726995 0 -537 0 -3.279101 0 -538 0 -4.533843 0 -539 0 -3.5858517 0 -540 0 -3.6101003 0 -541 0 -5.4949865 0 -542 0 -4.2921433 0 -543 0 -4.533843 0 -544 0 -4.589209 0 -545 0 -4.0598474 0 -546 1 11.390259 1 -547 0 -6.0596447 0 -548 0 -5.5856485 0 -549 1 6.3187485 1 -550 0 -5.0078387 0 -551 0 -5.4686823 0 -552 0 -3.1100621 0 -553 0 -1.7353668 0 -554 0 -5.4949865 0 -555 0 -1.9254999 0 -556 0 -3.4240565 0 -557 0 -3.93614 0 -558 0 -5.4818344 0 -559 0 -4.0598474 0 -560 0 -3.5726995 0 -561 0 -3.5726995 0 -562 0 -5.4686823 0 -563 0 -5.0078387 0 -564 0 -3.8496675 0 -565 1 12.086092 1 -566 0 -4.2270923 0 -567 0 -3.6343493 0 -568 1 4.1473055 1 -569 1 10.713882 1 -570 1 8.017664 1 -571 1 11.034658 1 -572 0 -5.0078387 0 -573 0 -5.95583 0 -574 1 5.950967 1 -575 0 -3.279101 0 -576 0 -4.0598474 0 -577 0 -5.95583 0 -578 0 -5.95583 0 -579 0 -5.4686823 0 -580 0 -3.7662487 0 -581 1 8.417797 1 -582 1 7.90798 1 -583 0 -5.4949865 0 -584 0 -2.9291954 0 -585 0 -6.442978 0 -586 1 13.98102 1 -587 0 -3.8628197 0 -588 1 5.463169 1 -589 0 -4.5469956 0 -590 1 3.9684038 1 -591 1 5.966527 1 -592 1 5.7801704 1 -593 0 -4.253397 0 -594 1 5.101776 1 -595 0 -4.0598474 0 -596 0 -4.2402444 0 -597 0 -2.9855018 0 -598 0 -5.0078387 0 -599 0 -3.6294346 0 -600 0 -5.0078387 0 -601 0 -6.0464916 0 -602 0 -4.533843 0 -603 1 4.8058825 1 -604 1 6.1928034 1 -605 1 9.95545 1 -606 0 -4.715564 0 -607 0 -6.442978 0 -608 1 11.148426 1 -609 0 -4.5469956 0 -610 1 8.926191 1 -611 1 6.9109592 1 -612 1 16.893515 1 -613 0 -5.226982 0 -614 0 -5.5724964 0 -615 0 -3.9466453 0 -616 0 -5.0078387 0 +412 1 8.967401 1 +413 0 -3.424508 0 +414 1 6.574872 1 +415 0 -0.8179264 0 +416 1 8.593532 1 +417 0 -6.1762524 0 +418 0 -2.0245438 0 +419 0 -5.7038784 0 +420 0 -2.7462745 0 +421 1 11.613648 1 +422 0 -3.0301971 0 +423 0 -3.6295166 0 +424 0 -5.668988 0 +425 1 14.596736 1 +426 0 -3.0489054 0 +427 1 4.065505 1 +428 0 -6.1762524 0 +429 0 -5.679842 0 +430 0 -5.7589035 0 +431 0 -2.7103744 0 +432 0 -3.9977875 0 +433 0 -4.6323185 0 +434 0 4.7780943 1 +435 1 7.255061 1 +436 1 3.4924908 1 +437 0 -5.172579 0 +438 0 -3.9676237 0 +439 0 -4.6761684 0 +440 1 10.246327 1 +441 0 -1.9451714 0 +442 0 -5.1030173 0 +443 0 -6.204343 0 +444 0 -2.501964 0 +445 0 -6.1653986 0 +446 0 -6.6618085 0 +447 0 -4.6761684 0 +448 0 -6.740807 0 +449 1 10.062795 1 +450 0 -4.122161 0 +451 0 -4.6761684 0 +452 0 -5.0251284 0 +453 1 8.612309 1 +454 0 -5.8618755 0 +455 1 0.8015022 1 +456 1 10.460095 1 +457 1 8.789279 1 +458 0 -4.395621 0 +459 0 -4.115073 0 +460 0 -4.0323086 0 +461 0 -3.8391252 0 +462 0 -3.5358982 0 +463 0 -4.9797544 0 +464 0 -5.172579 0 +465 1 9.567504 1 +466 1 9.408446 1 +467 1 7.540141 1 +468 0 -5.172579 0 +469 0 -5.8957047 0 +470 0 -5.2624307 0 +471 0 -3.5358982 0 +472 0 -4.2587566 0 +473 0 -5.172579 0 +474 0 -4.6761684 0 +475 0 -5.668988 0 +476 0 -4.8920307 0 +477 0 -5.172579 0 +478 0 -4.563906 0 +479 1 8.153737 1 +480 0 -4.755167 0 +481 0 -3.205307 0 +482 1 15.148742 1 +483 1 10.712245 1 +484 0 -4.395621 0 +485 0 -5.182016 0 +486 0 -5.2624307 0 +487 1 13.332547 1 +488 1 1.1748514 1 +489 1 -0.7666216 0 +490 0 -6.6618085 0 +491 1 6.429366 1 +492 0 -4.766021 0 +493 1 9.704236 1 +494 0 0.97516537 1 +495 0 -5.2624307 0 +496 0 -6.740807 0 +497 0 -5.060316 0 +498 0 -4.687022 0 +499 0 -4.687022 0 +500 0 -3.2086458 0 +501 0 -4.687022 0 +502 0 -4.3650923 0 +503 0 -4.2014656 0 +504 0 -6.6618085 0 +505 0 -5.4444637 0 +506 1 10.169904 1 +507 0 -5.289411 0 +508 0 -4.6761684 0 +509 0 -6.1653986 0 +510 0 -6.6618085 0 +511 0 -4.190612 0 +512 0 -4.6761684 0 +513 0 -5.2624307 0 +514 1 10.608193 1 +515 1 8.427064 1 +516 0 -6.740807 0 +517 0 -6.255251 0 +518 0 -5.066511 0 +519 1 6.2683277 1 +520 0 -6.58281 0 +521 0 -5.09358 0 +522 1 5.4384966 1 +523 1 7.66887 1 +524 0 -5.183432 0 +525 0 -5.3470583 0 +526 0 -5.172579 0 +527 0 -4.2014656 0 +528 0 -3.296733 0 +529 0 -4.766021 0 +530 1 6.384506 1 +531 0 -4.2942123 0 +532 0 -5.7696943 0 +533 0 -5.183432 0 +534 0 -5.679842 0 +535 0 -4.8076706 0 +536 0 -3.7050557 0 +537 0 -3.424508 0 +538 0 -4.687022 0 +539 0 -3.694202 0 +540 0 -3.7331467 0 +541 0 -5.668988 0 +542 0 -4.451644 0 +543 0 -4.687022 0 +544 0 -4.796817 0 +545 0 -4.190612 0 +546 1 11.1768465 1 +547 0 -6.244397 0 +548 0 -5.747987 0 +549 1 6.1616106 1 +550 0 -5.183432 0 +551 0 -5.690696 0 +552 0 -3.3047552 0 +553 0 -1.9243431 0 +554 0 -5.668988 0 +555 0 -2.141615 0 +556 0 -3.5548706 0 +557 0 -4.0323086 0 +558 0 -5.679842 0 +559 0 -4.190612 0 +560 0 -3.7050557 0 +561 0 -3.7050557 0 +562 0 -5.690696 0 +563 0 -5.183432 0 +564 0 -4.0086412 0 +565 1 11.879315 1 +566 0 -4.417328 0 +567 0 -3.772091 0 +568 1 3.9970121 1 +569 1 10.578958 1 +570 1 7.813756 1 +571 1 10.822152 1 +572 0 -5.183432 0 +573 0 -6.1762524 0 +574 1 5.7168713 1 +575 0 -3.424508 0 +576 0 -4.190612 0 +577 0 -6.1762524 0 +578 0 -6.1762524 0 +579 0 -5.690696 0 +580 0 -3.9100647 0 +581 1 8.170806 1 +582 1 7.6789007 1 +583 0 -5.668988 0 +584 0 -2.9875202 0 +585 0 -6.6618085 0 +586 1 13.784331 1 +587 0 -3.9977875 0 +588 1 5.322174 1 +589 0 -4.6761684 0 +590 1 3.6173382 1 +591 1 5.7292347 1 +592 1 5.8126287 1 +593 0 -4.395621 0 +594 1 4.848382 1 +595 0 -4.190612 0 +596 0 -4.4064746 0 +597 0 -3.14396 0 +598 0 -5.183432 0 +599 0 -3.801262 0 +600 0 -5.183432 0 +601 0 -6.255251 0 +602 0 -4.687022 0 +603 1 4.6091757 1 +604 1 6.150976 1 +605 1 9.757047 1 +606 0 -4.850649 0 +607 0 -6.6618085 0 +608 1 10.875676 1 +609 0 -4.6761684 0 +610 1 9.026143 1 +611 1 6.6211424 1 +612 1 16.623665 1 +613 0 -5.4553175 0 +614 0 -5.7588406 0 +615 0 -4.125927 0 +616 0 -5.183432 0 617 0 ? 0 -618 0 -4.533843 0 -619 0 -4.0598474 0 -620 0 -5.0078387 0 -621 0 0.3560543 1 -622 0 -2.2074018 0 -623 0 -6.442978 0 -624 0 -3.8450818 0 -625 0 -3.4678864 0 -626 1 5.760166 1 -627 0 -4.1699953 0 -628 0 -5.9689827 0 -629 0 -5.0209913 0 -630 0 -3.358376 0 -631 0 -4.0598474 0 -632 0 -6.442978 0 -633 1 4.3299494 1 -634 0 -5.4949865 0 -635 0 -4.6141906 0 -636 1 10.12752 1 -637 0 -2.4650183 0 -638 0 -5.0209913 0 -639 0 -3.93614 0 -640 0 -4.4101357 0 -641 0 -5.0078387 0 -642 0 -5.0078387 0 -643 0 -6.442978 0 -644 0 -5.9689827 0 -645 0 -5.0078387 0 -646 0 -6.021953 0 -647 0 -5.832123 0 -648 1 12.2362 1 -649 0 -5.0078387 0 -650 0 -3.7496572 0 -651 0 -5.2175484 0 -652 0 -3.8628197 0 -653 0 -4.533843 0 -654 0 -4.520691 0 -655 0 -5.0078387 0 -656 0 -4.0598474 0 -657 0 -0.4869156 0 -658 1 9.086258 1 -659 0 -6.442978 0 -660 0 -5.95583 0 -661 0 -4.046695 0 -662 0 -5.3686323 0 -663 0 -5.3686323 0 -664 0 -4.3969836 0 -665 0 -6.442978 0 -666 0 -3.496932 0 -667 0 -4.520691 0 -668 1 2.804366 1 -669 1 8.147337 1 -670 1 6.4856215 1 -671 0 -4.087837 0 -672 0 -4.9946866 0 -673 0 -3.9078827 0 -674 0 -5.95583 0 -675 0 -4.140195 0 -676 0 -5.6622314 0 -677 0 -4.5469956 0 -678 0 -6.442978 0 -679 0 -5.9689827 0 -680 1 16.78001 1 -681 1 9.801079 1 -682 0 -3.3756719 0 -683 0 -6.442978 0 -684 0 -6.442978 0 -685 0 -6.442978 0 -686 0 -6.442978 0 -687 0 -4.613783 0 -688 0 -5.0209913 0 -689 0 -4.177704 0 -690 0 -5.832123 0 -691 1 4.4967804 1 -692 0 -5.4949865 0 -693 0 -4.684228 0 -694 0 -4.9545784 0 -695 0 -5.9689827 0 -696 1 6.7127876 1 -697 1 5.064643 1 -698 1 6.1036224 1 +618 0 -4.687022 0 +619 0 -4.190612 0 +620 0 -5.183432 0 +621 0 0.21396732 1 +622 0 -2.2733846 0 +623 0 -6.6618085 0 +624 0 -3.9476805 0 +625 0 -3.656497 0 +626 1 5.46276 1 +627 0 -4.335535 0 +628 0 -6.1653986 0 +629 0 -5.172579 0 +630 0 -3.506401 0 +631 0 -4.190612 0 +632 0 -6.6618085 0 +633 1 4.123639 1 +634 0 -5.668988 0 +635 0 -4.827346 0 +636 1 9.884419 1 +637 0 -2.6317916 0 +638 0 -5.172579 0 +639 0 -4.0323086 0 +640 0 -4.528719 0 +641 0 -5.183432 0 +642 0 -5.183432 0 +643 0 -6.6618085 0 +644 0 -6.1653986 0 +645 0 -5.183432 0 +646 0 -6.268999 0 +647 0 -6.0179486 0 +648 1 11.790706 1 +649 0 -5.183432 0 +650 0 -3.8956623 0 +651 0 -5.4201303 0 +652 0 -3.9977875 0 +653 0 -4.687022 0 +654 0 -4.697876 0 +655 0 -5.183432 0 +656 0 -4.190612 0 +657 0 -0.8079653 0 +658 1 8.931214 1 +659 0 -6.6618085 0 +660 0 -6.1762524 0 +661 0 -4.2014656 0 +662 0 -5.615157 0 +663 0 -5.615157 0 +664 0 -4.5395722 0 +665 0 -6.6618085 0 +666 0 -3.645947 0 +667 0 -4.697876 0 +668 1 2.434762 1 +669 1 7.828244 1 +670 1 6.2092333 1 +671 0 -4.2636833 0 +672 0 -5.194286 0 +673 0 -4.0920978 0 +674 0 -6.1762524 0 +675 0 -4.3309355 0 +676 0 -5.8957047 0 +677 0 -4.6761684 0 +678 0 -6.6618085 0 +679 0 -6.1653986 0 +680 1 16.591042 1 +681 1 9.56899 1 +682 0 -3.5122313 0 +683 0 -6.6618085 0 +684 0 -6.6618085 0 +685 0 -6.6618085 0 +686 0 -6.6618085 0 +687 0 -4.7977834 0 +688 0 -5.172579 0 +689 0 -4.4534364 0 +690 0 -6.0179486 0 +691 1 4.169842 1 +692 0 -5.668988 0 +693 0 -4.8679504 0 +694 0 -5.132524 0 +695 0 -6.1653986 0 +696 1 6.262699 1 +697 1 4.767309 1 +698 1 5.8011494 1 diff --git a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-def-CV-breast-cancer-out.txt b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-def-CV-breast-cancer-out.txt index 703a4d879b..c3c9d949bb 100644 --- a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-def-CV-breast-cancer-out.txt +++ b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-def-CV-breast-cancer-out.txt @@ -11,43 +11,43 @@ Confusion table ||====================== PREDICTED || positive | negative | Recall TRUTH ||====================== - positive || 133 | 1 | 0.9925 - negative || 10 | 210 | 0.9545 + positive || 134 | 0 | 1.0000 + negative || 9 | 211 | 0.9591 ||====================== -Precision || 0.9301 | 0.9953 | -OVERALL 0/1 ACCURACY: 0.968927 -LOG LOSS/instance: 0.138800 +Precision || 0.9371 | 1.0000 | +OVERALL 0/1 ACCURACY: 0.974576 +LOG LOSS/instance: 0.117268 Test-set entropy (prior Log-Loss/instance): 0.956998 -LOG-LOSS REDUCTION (RIG): 0.854963 -AUC: 0.994267 +LOG-LOSS REDUCTION (RIG): 0.877462 +AUC: 0.994607 Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. TEST POSITIVE RATIO: 0.3191 (105.0/(105.0+224.0)) Confusion table ||====================== PREDICTED || positive | negative | Recall TRUTH ||====================== - positive || 101 | 4 | 0.9619 - negative || 4 | 220 | 0.9821 + positive || 99 | 6 | 0.9429 + negative || 3 | 221 | 0.9866 ||====================== -Precision || 0.9619 | 0.9821 | -OVERALL 0/1 ACCURACY: 0.975684 -LOG LOSS/instance: 0.157194 +Precision || 0.9706 | 0.9736 | +OVERALL 0/1 ACCURACY: 0.972644 +LOG LOSS/instance: 0.176988 Test-set entropy (prior Log-Loss/instance): 0.903454 -LOG-LOSS REDUCTION (RIG): 0.826008 -AUC: 0.982185 +LOG-LOSS REDUCTION (RIG): 0.804098 +AUC: 0.984396 OVERALL RESULTS --------------------------------------- -AUC: 0.988226 (0.0060) -Accuracy: 0.972305 (0.0034) -Positive precision: 0.945987 (0.0159) -Positive recall: 0.977221 (0.0153) -Negative precision: 0.988702 (0.0066) -Negative recall: 0.968344 (0.0138) -Log-loss: 0.147997 (0.0092) -Log-loss reduction: 0.840485 (0.0145) -F1 Score: 0.961097 (0.0008) -AUPRC: 0.986066 (0.0034) +AUC: 0.989501 (0.0051) +Accuracy: 0.973610 (0.0010) +Positive precision: 0.953826 (0.0168) +Positive recall: 0.971429 (0.0286) +Negative precision: 0.986784 (0.0132) +Negative recall: 0.972849 (0.0138) +Log-loss: 0.147128 (0.0299) +Log-loss reduction: 0.840780 (0.0367) +F1 Score: 0.962015 (0.0055) +AUPRC: 0.986587 (0.0035) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-def-CV-breast-cancer.txt b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-def-CV-breast-cancer.txt index 1055c7a08f..137b8873b1 100644 --- a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-def-CV-breast-cancer.txt +++ b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-def-CV-breast-cancer.txt @@ -1,700 +1,700 @@ Instance Label Score Probability Log-loss Assigned -5 1 3.192721 0.9995258 0.00068430925266268561 1 -6 0 0.31271142 0.5096692 1.0281726628467021 1 -8 0 -1.3462902 0.012771749 0.018544414860777944 0 -9 0 -1.4529887 0.009662693 0.01400810734613289 0 -10 0 -1.4078557 0.010874045 0.015773849973586711 0 -11 0 -1.4552101 0.009606653 0.013926471498510302 0 -18 1 1.9908078 0.9887438 0.016331378013210652 1 -20 1 1.9282161 0.9867448 0.019251052731376804 1 -21 1 2.2146642 0.99374056 0.0090588463608975659 1 -25 1 0.77787775 0.78049725 0.35753454042303912 1 -28 0 -1.4552101 0.009606653 0.013926471498510302 0 -31 0 -1.4188342 0.010566232 0.015324957073613255 0 -32 1 2.3759427 0.9959046 0.0059205096261509705 1 -35 0 -1.4552101 0.009606653 0.013926471498510302 0 -37 0 -1.1775297 0.019811774 0.02886927717583929 0 +5 1 2.3422444 0.9976212 0.0034360022379797063 1 +6 0 -0.8979866 0.022918249 0.033448818795447796 0 +8 0 -1.1725245 0.010128294 0.014686540686377271 0 +9 0 -1.2562976 0.007880984 0.011414895973673126 0 +10 0 -1.2001765 0.0093239425 0.013514709124084631 0 +11 0 -1.2384154 0.008314926 0.012046053477396252 0 +18 1 1.5604059 0.97530615 0.036072936011745257 1 +20 1 1.5261991 0.97269046 0.039947322200609377 1 +21 1 1.7936352 0.9876415 0.017940616606305956 1 +25 1 0.8390706 0.8170441 0.29151418141546054 1 +28 0 -1.2384154 0.008314926 0.012046053477396252 0 +31 0 -1.232281 0.00846918 0.012270478382962932 0 +32 1 1.9197828 0.9915254 0.012278347417683086 1 +35 0 -1.2384154 0.008314926 0.012046053477396252 0 +37 0 -1.1595228 0.010529869 0.015271937740027402 0 40 0 ? ? ? 0 -41 1 1.2809899 0.930783 0.1034832742252651 1 -44 1 2.5676048 0.99752873 0.003569699004464264 1 -45 0 -1.4192019 0.010556074 0.015310145847072318 0 -46 1 1.4376967 0.9531627 0.069205554049795295 1 -48 0 -1.4035003 0.010998605 0.015955538696674449 0 -50 1 1.0968927 0.89206713 0.1647758076195768 1 -51 1 0.5286348 0.64784414 0.62628133673439668 1 -52 1 1.5422581 0.96406907 0.05279158628001978 1 -54 1 2.2490754 0.99428177 0.0082733408427395384 1 -56 1 1.6056342 0.969443 0.044771985163227578 1 -60 1 0.9972039 0.86394596 0.21098701888472435 1 -63 1 0.74569666 0.7655731 0.38538798574135286 1 -64 0 -1.4530833 0.009660303 0.01400462463965267 0 -66 0 -1.4115138 0.010770508 0.015622843432160721 0 -68 1 2.9430315 0.9990827 0.0013240138604818531 1 -69 0 -1.4171377 0.010613227 0.015393482872561901 0 -70 0 -1.3868607 0.011487602 0.016669034799060203 0 -71 1 2.6230502 0.99786496 0.0030835021472300035 1 -72 0 -1.1639595 0.020520657 0.029913028632274973 0 -73 1 2.121044 0.9919966 0.011592938895840786 1 -74 1 1.0584506 0.88188386 0.1813394234154109 1 -76 0 -1.3763652 0.01180703 0.017135302488555097 0 -77 0 -1.4423684 0.009935132 0.014405043081600288 0 -79 0 -1.4663328 0.009330851 0.013524769896968257 0 -82 0 -1.4030102 0.011012708 0.015976111467283581 0 -88 0 -1.4115138 0.010770508 0.015622843432160721 0 -90 0 -1.4194673 0.010548745 0.015299460185409182 0 -91 0 -1.488508 0.008804185 0.012757999163627449 0 -92 0 -1.4115138 0.010770508 0.015622843432160721 0 -93 0 -1.4530833 0.009660303 0.01400462463965267 0 -95 0 -1.4194673 0.010548745 0.015299460185409182 0 -96 0 -1.4862804 0.00885573 0.012833025327400363 0 -97 0 -1.4049767 0.010956224 0.0158937175406033 0 -98 1 2.6602142 0.9980644 0.0027951883606189454 1 -99 1 2.6854086 0.9981889 0.0026152153301300263 1 -100 1 1.7509568 0.97898656 0.030639039022914153 1 -102 0 -1.3805804 0.0116777 0.016946501812626372 0 -104 1 3.5134978 0.99979687 0.0002930882059670945 1 -105 1 1.1370488 0.90187234 0.14900486540615213 1 -106 1 2.133229 0.99224836 0.011226827482112123 1 -108 0 -1.4496208 0.009748278 0.014132790248751441 0 -109 1 1.914584 0.98626506 0.019952665224579537 1 -111 1 1.2717487 0.9291921 0.10595116644163287 1 -112 1 2.1944647 0.99339944 0.0095541595852411017 1 -113 1 2.886482 0.9989349 0.0015373978618641254 1 -115 0 -1.2772555 0.015290121 0.022229362427259085 0 -117 1 2.519053 0.9971912 0.0040579569047968205 1 -120 0 -1.3902658 0.011385816 0.016520489619359105 0 -121 0 -1.3535905 0.0125306565 0.018192135279811766 0 -122 1 2.8106294 0.9986987 0.0018785851728224208 1 -123 1 1.547194 0.96451837 0.052119381732274699 1 -125 0 -1.4530833 0.009660303 0.01400462463965267 0 -128 1 1.3565444 0.94259703 0.085286956831716013 1 -129 0 -1.5726374 0.007060734 0.010222618271006645 0 -131 0 -1.4188342 0.010566232 0.015324957073613255 0 -132 1 2.2541509 0.9943576 0.0081633350081040026 1 -133 0 -1.4524355 0.009676701 0.014028513977170701 0 -137 0 -1.4908748 0.008749744 0.012678761369425539 0 -138 0 -1.4239863 0.010424764 0.015118697237142804 0 -141 0 -1.4881064 0.008813457 0.012771493716677452 0 -144 0 -1.4552101 0.009606653 0.013926471498510302 0 +41 1 1.2136381 0.9326625 0.10057300350215044 1 +44 1 2.1646152 0.995938 0.0058721572698145451 1 +45 0 -1.2113214 0.009017945 0.013069162382252208 0 +46 1 1.2518405 0.93956035 0.089942257065595443 1 +48 0 -1.2516923 0.007990543 0.011574220453025332 0 +50 1 1.1002063 0.90767354 0.13975459713507837 1 +51 1 0.68538195 0.73730874 0.43965923501734078 1 +52 1 1.4014324 0.9606762 0.057877858614391035 1 +54 1 1.7354605 0.9853012 0.021363285666455459 1 +56 1 1.4347819 0.9643114 0.052428959990264194 1 +60 1 1.0599462 0.89696014 0.15688422139125702 1 +63 1 0.86943996 0.83036536 0.268181834610912 1 +64 0 -1.2272866 0.008596861 0.012456267896939406 0 +66 0 -1.2479138 0.008081559 0.011706592959255056 0 +68 1 2.3008437 0.99730504 0.0038932602464770804 1 +69 0 -1.2118624 0.009003345 0.013047907013782868 0 +70 0 -1.2067921 0.009141082 0.013248438798178453 0 +71 1 2.1951663 0.9962949 0.0053552337139906514 1 +72 0 -1.0938455 0.012811873 0.01860305187618028 0 +73 1 1.6421516 0.9806062 0.028254214108077248 1 +74 1 1.0454789 0.89284897 0.16351194061092869 1 +76 0 -1.1468637 0.010935987 0.015864199029258574 0 +77 0 -1.2640618 0.0076996386 0.011151215902728351 0 +79 0 -1.2611338 0.0077675357 0.011249934218045958 0 +82 0 -1.2521169 0.007980378 0.011559438201725308 0 +88 0 -1.2479138 0.008081559 0.011706592959255056 0 +90 0 -1.2229315 0.008709754 0.012620559823242145 0 +91 0 -1.2556949 0.007895238 0.011435623493958765 0 +92 0 -1.2479138 0.008081559 0.011706592959255056 0 +93 0 -1.2272866 0.008596861 0.012456267896939406 0 +95 0 -1.2229315 0.008709754 0.012620559823242145 0 +96 0 -1.2439907 0.008177149 0.011845630489222336 0 +97 0 -1.2542456 0.007929614 0.011485613187028508 0 +98 1 2.1175616 0.99532026 0.0067672838755976307 1 +99 1 2.0379202 0.99405456 0.0086030626608249762 1 +100 1 1.4803522 0.9687591 0.045790108595457714 1 +102 0 -1.218076 0.0088373525 0.012806275136592609 0 +104 1 2.8564382 0.9994961 0.0007271538861604815 1 +105 1 1.1049452 0.90886664 0.13785946855375081 1 +106 1 1.656333 0.98140466 0.02707996998711153 1 +108 0 -1.2055137 0.009176139 0.013299482624196092 0 +109 1 1.576021 0.9764175 0.03442996934675914 1 +111 1 1.0971423 0.90689474 0.14099297734786179 1 +112 1 1.7211893 0.98466337 0.022297508771260729 1 +113 1 2.1748955 0.99606174 0.0056929223153520947 1 +115 0 -1.0728248 0.013640635 0.019814728145547817 0 +117 1 1.9577936 0.992438 0.010951091827503274 1 +120 0 -1.1973941 0.009401926 0.013628278681614545 0 +121 0 -1.2057146 0.009170622 0.01329144935994855 0 +122 1 2.076021 0.9946977 0.0076699686209208079 1 +123 1 1.2354816 0.9366914 0.094354270194486917 1 +125 0 -1.2272866 0.008596861 0.012456267896939406 0 +128 1 1.1640284 0.92261463 0.11619991901837462 1 +129 0 -1.4118875 0.0049394183 0.0071437317661628402 0 +131 0 -1.232281 0.00846918 0.012270478382962932 0 +132 1 1.7101815 0.98415285 0.023045690050001347 1 +133 0 -1.2373763 0.008340856 0.01208377654478787 0 +137 0 -1.2532576 0.007953135 0.011519819156242013 0 +138 0 -1.2288303 0.008557193 0.012398544228788652 0 +141 0 -1.2414014 0.008240851 0.01193829304157505 0 +144 0 -1.2384154 0.008314926 0.012046053477396252 0 145 0 ? ? ? 0 -147 0 -1.4122752 0.01074908 0.015591593305739363 0 -150 0 -1.4078557 0.010874045 0.015773849973586711 0 -151 1 1.6789396 0.9746906 0.036983740030651899 1 -152 1 2.5298893 0.9972703 0.0039435295365728758 1 -154 0 -1.4831115 0.008929576 0.012940517754445463 0 -156 0 -1.4304588 0.010249696 0.014863489132720689 0 -161 0 -1.407595 0.010881461 0.015784666838719357 0 +147 0 -1.2099732 0.009054424 0.013122270310084439 0 +150 0 -1.2001765 0.0093239425 0.013514709124084631 0 +151 1 1.348086 0.9541131 0.06776776329340771 1 +152 1 2.07214 0.99463546 0.0077602249398833886 1 +154 0 -1.228457 0.00856677 0.012412479933523967 0 +156 0 -1.2004136 0.009317326 0.013505074284140927 0 +161 0 -1.2366272 0.008359601 0.012111047189131965 0 164 0 ? ? ? 0 -167 1 2.4332309 0.99647826 0.0050897651488173556 1 -169 0 -1.4668974 0.009317063 0.013504690465132399 0 -171 0 -1.4194673 0.010548745 0.015299460185409182 0 -173 1 3.952484 0.99993634 9.1841658808920751E-05 1 -174 1 1.7119659 0.9767577 0.033927363647315606 1 -176 0 -1.4188342 0.010566232 0.015324957073613255 0 -177 1 1.8848073 0.9851567 0.021574853720767784 1 -179 1 1.0500653 0.87955487 0.18515451724266119 1 -180 0 -1.4078557 0.010874045 0.015773849973586711 0 -181 0 -1.4831115 0.008929576 0.012940517754445463 0 -183 1 2.3670325 0.9958074 0.0060613450978592278 1 -187 1 3.4790509 0.9997775 0.0003210413356730718 1 -188 1 2.36301 0.9957628 0.0061259388644517325 1 -189 0 -1.3655348 0.012145848 0.017630038642071411 0 -191 1 2.833822 0.998776 0.0017669134230232214 1 -192 0 -1.4490774 0.009762159 0.014153012753613161 0 -196 0 1.9304646 0.98682237 6.2457649054561086 1 -198 0 -1.4831115 0.008929576 0.012940517754445463 0 -199 0 -1.4552015 0.009606869 0.013926786240718135 0 -201 1 2.8761835 0.99890554 0.0015798374101686152 1 -202 0 -1.4194673 0.010548745 0.015299460185409182 0 -204 0 -1.4194673 0.010548745 0.015299460185409182 0 -205 1 3.0781524 0.9993581 0.00092633791539752488 1 -206 1 1.8323022 0.9829839 0.024760326392249166 1 -207 0 -1.4078557 0.010874045 0.015773849973586711 0 -209 0 -1.3985585 0.011141644 0.016164210433150503 0 -210 1 3.5384111 0.9998098 0.00027442441782451788 1 -211 1 2.4245603 0.9963969 0.0052075629534229758 1 -212 0 -1.4194673 0.010548745 0.015299460185409182 0 -216 0 -1.4530833 0.009660303 0.01400462463965267 0 -218 1 2.3203871 0.99525976 0.0068549781098011502 1 -219 0 -1.3283615 0.013383448 0.019438604119709272 0 -223 1 1.7294853 0.9777864 0.032408723671712908 1 -226 1 2.5808032 0.99761325 0.003447466400898522 1 -228 0 -1.4078557 0.010874045 0.015773849973586711 0 -233 1 1.7235589 0.9774435 0.032914759397390546 1 -237 1 1.9862187 0.98860794 0.016529596902120164 1 -239 1 1.6215922 0.97066844 0.042949516781757888 1 -240 0 -1.3102634 0.014030224 0.020384671679905413 0 -241 0 -1.4008535 0.011074988 0.016066966384989452 0 -242 0 -1.4188342 0.010566232 0.015324957073613255 0 -244 0 -1.4194673 0.010548745 0.015299460185409182 0 -246 1 2.8064017 0.9986841 0.0018996806528914184 1 -247 1 1.0271662 0.8729921 0.19595949916508731 1 -248 0 -1.3904947 0.011379007 0.0165105533553958 0 +167 1 2.072563 0.9946423 0.0077502826356724876 1 +169 0 -1.20626 0.009155659 0.013269663220583924 0 +171 0 -1.2229315 0.008709754 0.012620559823242145 0 +173 1 2.9003725 0.9995587 0.00063682030230397479 1 +174 1 1.3856966 0.95883995 0.0606380704270615 1 +176 0 -1.232281 0.00846918 0.012270478382962932 0 +177 1 1.6888264 0.98311436 0.024568845613762765 1 +179 1 0.9162498 0.8493702 0.23553463332929456 1 +180 0 -1.2001765 0.0093239425 0.013514709124084631 0 +181 0 -1.228457 0.00856677 0.012412479933523967 0 +183 1 1.8139207 0.9883678 0.016880090886474304 1 +187 1 2.615671 0.99895746 0.0015048588509106512 1 +188 1 1.8776768 0.9903866 0.013936292301738214 1 +189 0 -1.1660835 0.010325302 0.014973700238501147 0 +191 1 2.057691 0.9943974 0.0081055680073918602 1 +192 0 -1.2655382 0.0076656267 0.011101767206480718 0 +196 0 1.6092172 0.9786202 5.547605912530484 1 +198 0 -1.228457 0.00856677 0.012412479933523967 0 +199 0 -1.2484847 0.008067742 0.011686496825873889 0 +201 1 2.114022 0.9952702 0.0068398580339468626 1 +202 0 -1.2229315 0.008709754 0.012620559823242145 0 +204 0 -1.2229315 0.008709754 0.012620559823242145 0 +205 1 2.2825005 0.99715185 0.0041148721631914234 1 +206 1 1.5528458 0.9747499 0.036895959596087165 1 +207 0 -1.2001765 0.0093239425 0.013514709124084631 0 +209 0 -1.2342625 0.008419046 0.012197534325077079 0 +210 1 2.608253 0.99893385 0.0015389473568821269 1 +211 1 1.9089165 0.991245 0.012686452352322526 1 +212 0 -1.2229315 0.008709754 0.012620559823242145 0 +216 0 -1.2272866 0.008596861 0.012456267896939406 0 +218 1 1.938983 0.99199927 0.011589038071517037 1 +219 0 -1.1974437 0.009400531 0.013626246845397564 0 +223 1 1.4449766 0.9653566 0.050866143970474634 1 +226 1 1.9788362 0.9929006 0.010278784666155427 1 +228 0 -1.2001765 0.0093239425 0.013514709124084631 0 +233 1 1.3701234 0.9569421 0.063496486091890819 1 +237 1 1.7527282 0.9860381 0.020284718609316355 1 +239 1 1.3340445 0.95221925 0.070634303578795812 1 +240 0 -1.1520904 0.010766463 0.0156169445929962 0 +241 0 -1.223617 0.008691888 0.012594559038885133 0 +242 0 -1.232281 0.00846918 0.012270478382962932 0 +244 0 -1.2229315 0.008709754 0.012620559823242145 0 +246 1 2.08892 0.99489933 0.0073775388960712592 1 +247 1 0.9662375 0.86769426 0.20474131163513853 1 +248 0 -1.1930147 0.009525983 0.013808964851621683 0 249 0 ? ? ? 0 -250 0 -1.4637281 0.009394727 0.013617794003413085 0 -252 0 1.2930615 0.9328112 3.8956354392696553 1 -254 1 2.0918605 0.9913602 0.012518772740002581 1 -257 0 -1.4552015 0.009606869 0.013926786240718135 0 -258 0 -1.416158 0.010640464 0.015433198933891867 0 -259 0 1.2175665 0.9191675 3.6289210537828396 1 -260 1 2.3240974 0.9953058 0.0067882781675344759 1 -262 1 2.5283754 0.9972594 0.0039593091089041993 1 -267 1 1.3127309 0.93599814 0.095422429254134244 1 -268 1 2.7037425 0.99827445 0.0024915991850600612 1 -269 0 -1.4194673 0.010548745 0.015299460185409182 0 -271 0 -1.4049767 0.010956224 0.0158937175406033 0 -272 1 1.3127309 0.93599814 0.095422429254134244 1 +250 0 -1.2034192 0.009233863 0.013383534893272504 0 +252 0 1.1310676 0.91519725 3.5597451951363643 1 +254 1 1.6307943 0.9799427 0.029230731941557696 1 +257 0 -1.2484847 0.008067742 0.011686496825873889 0 +258 0 -1.2406003 0.008260659 0.011967108123699813 0 +259 0 1.0986385 0.9072758 3.4309102109633094 1 +260 1 1.780022 0.9871292 0.018689154600506006 1 +262 1 1.9265935 0.9916966 0.012029290997083615 1 +267 1 1.1061206 0.90916044 0.1373931924992165 1 +268 1 2.280658 0.997136 0.0041378113718746178 1 +269 0 -1.2229315 0.008709754 0.012620559823242145 0 +271 0 -1.2542456 0.007929614 0.011485613187028508 0 +272 1 1.1061206 0.90916044 0.1373931924992165 1 275 0 ? ? ? 0 -276 0 -1.4552015 0.009606869 0.013926786240718135 0 -277 0 -1.4530833 0.009660303 0.01400462463965267 0 -278 0 -1.4194673 0.010548745 0.015299460185409182 0 -279 1 1.990224 0.9887266 0.016356425671825971 1 -280 0 -1.416158 0.010640464 0.015433198933891867 0 -283 1 1.6401696 0.972035 0.040919846769262407 1 -284 1 1.6578851 0.9732805 0.039072460060008155 1 -285 1 3.6651192 0.9998639 0.00019633155440699929 1 -288 1 1.0698746 0.88499403 0.17626037180495394 1 -290 0 -1.4831115 0.008929576 0.012940517754445463 0 -291 0 -1.4194673 0.010548745 0.015299460185409182 0 -293 1 1.3049293 0.9347513 0.097345565286826974 1 -296 0 0.96243674 0.8527751 2.7639063708445106 1 +276 0 -1.2484847 0.008067742 0.011686496825873889 0 +277 0 -1.2272866 0.008596861 0.012456267896939406 0 +278 0 -1.2229315 0.008709754 0.012620559823242145 0 +279 1 1.6634443 0.98179275 0.026509584630759111 1 +280 0 -1.2406003 0.008260659 0.011967108123699813 0 +283 1 1.3519057 0.95461583 0.067007832604735862 1 +284 1 1.3838221 0.95861584 0.060975316674272396 1 +285 1 2.6684916 0.9991111 0.0012829589215421731 1 +288 1 1.0538921 0.895257 0.15962620708478997 1 +290 0 -1.228457 0.00856677 0.012412479933523967 0 +291 0 -1.2229315 0.008709754 0.012620559823242145 0 +293 1 1.1394588 0.9171446 0.12477888852758001 1 +296 0 0.9756187 0.8709149 2.9536053485661498 1 297 0 ? ? ? 0 -299 1 2.0153353 0.9894431 0.015311316399646006 1 -300 1 2.1153967 0.9918772 0.011766579604628905 1 -301 0 -1.4194673 0.010548745 0.015299460185409182 0 -303 0 -1.4194673 0.010548745 0.015299460185409182 0 -304 1 1.4359963 0.9529616 0.069509977769646764 1 -308 1 2.2193007 0.9938163 0.0089488671436337795 1 -309 0 -1.1012988 0.024128996 0.035237637623780134 0 -311 0 -1.4831115 0.008929576 0.012940517754445463 0 -312 1 1.4203453 0.95107156 0.072374198674943963 1 -314 0 -1.4396616 0.010005775 0.014507985306325046 0 -316 1 1.312181 0.935911 0.095556751284067876 1 -317 1 2.4755049 0.9968495 0.0045524185087478346 1 -319 0 1.3291107 0.93854403 4.0243031414384935 1 +299 1 1.5977252 0.97788143 0.032268546317352535 1 +300 1 1.795045 0.98769337 0.017864870003746217 1 +301 0 -1.2229315 0.008709754 0.012620559823242145 0 +303 0 -1.2229315 0.008709754 0.012620559823242145 0 +304 1 1.2114133 0.932239 0.10122823208472417 1 +308 1 1.9096477 0.99126416 0.012658518855034488 1 +309 0 -1.0598819 0.014176994 0.020599444992625792 0 +311 0 -1.228457 0.00856677 0.012412479933523967 0 +312 1 1.2908002 0.9459099 0.080225298187381611 1 +314 0 -1.2033671 0.009235303 0.013385631482353349 0 +316 1 1.1155818 0.91149414 0.13369472024420964 1 +317 1 1.9351709 0.99190736 0.011722712330300641 1 +319 0 1.1727912 0.92448413 3.7270763956938961 1 321 0 ? ? ? 0 -323 1 1.4047902 0.94912183 0.075334805993493639 1 -327 0 -1.4530833 0.009660303 0.01400462463965267 0 -328 1 1.2504702 0.9254 0.11185096684337191 1 -329 1 2.1595812 0.9927663 0.01047392157809102 1 -331 0 -1.3625069 0.01224228 0.017770877917333231 0 -332 0 -1.4078642 0.010873804 0.015773498151757607 0 -333 1 1.4430139 0.9537864 0.068261923966845606 1 -336 1 1.4525254 0.9548823 0.066605143324864419 1 -338 0 -1.4396616 0.010005775 0.014507985306325046 0 -343 0 -1.4831115 0.008929576 0.012940517754445463 0 -344 1 2.8907647 0.9989469 0.0015200952805456086 1 -346 0 -1.454973 0.00961262 0.013935163563942931 0 -347 0 -1.4665179 0.009326328 0.013518182510513826 0 -348 1 -1.2562517 0.016149154 5.9523976068209432 0 -349 1 1.368182 0.9442394 0.082775445428497504 1 -350 0 -1.483142 0.0089288615 0.012939477917220827 0 -352 0 0.9176493 0.8372787 2.6195251949431575 1 -353 1 2.3254437 0.9953224 0.0067641736361085097 1 -354 0 -1.4530833 0.009660303 0.01400462463965267 0 -355 0 -1.457952 0.009537922 0.013826355758292658 0 -358 1 1.8535075 0.98389673 0.023421193523213833 1 -360 1 4.076732 0.99995416 6.6128844808261867E-05 1 -361 1 1.8173494 0.9823098 0.025749975862691399 1 -366 1 3.5200806 0.9998004 0.00028801369589076065 1 -368 0 -1.437181 0.010070953 0.014602970314783701 0 -370 0 -1.4284434 0.010303893 0.0149424913218754 0 -371 0 -1.437181 0.010070953 0.014602970314783701 0 -373 0 -1.4451452 0.009863174 0.014300190761160224 0 -376 0 -1.4530833 0.009660303 0.01400462463965267 0 -377 0 -1.4396616 0.010005775 0.014507985306325046 0 -378 0 -1.4514455 0.009701818 0.014065104450139877 0 -379 0 -1.3010159 0.014372502 0.020885589239682686 0 -381 1 2.4906044 0.99697244 0.0043744687100195252 1 -383 0 -1.4881064 0.008813457 0.012771493716677452 0 -384 0 -1.4881064 0.008813457 0.012771493716677452 0 -387 0 -1.329933 0.013328694 0.019358541009786727 0 -388 0 -1.4529943 0.009662552 0.014007901124089213 0 -389 0 -1.3949994 0.011245799 0.016316175959814438 0 -391 1 2.4490592 0.9966221 0.0048815497730760786 1 -392 0 -1.4552015 0.009606869 0.013926786240718135 0 -395 0 -1.4552015 0.009606869 0.013926786240718135 0 -396 0 -1.416158 0.010640464 0.015433198933891867 0 -398 0 -1.452887 0.009665267 0.014011855979604353 0 -399 0 -1.4394311 0.0100118155 0.014516788095020891 0 -404 0 -1.4821421 0.008952287 0.012973578732603224 0 -406 0 -1.4232016 0.010446191 0.01514993576396167 0 -409 0 -1.4275737 0.010327371 0.014976715543268503 0 -413 0 -1.3726467 0.011922293 0.017303588906559128 0 -414 1 1.8375838 0.98321587 0.024419894816904819 1 -415 0 -1.2291577 0.017327672 0.025217664476067117 0 -416 1 2.6160972 0.99782544 0.0031406375111272277 1 -418 0 -1.256744 0.016128484 0.023458169122501132 0 -419 0 -1.5255483 0.007989398 0.011572555850730927 0 -422 0 -1.3332301 0.013214536 0.019191630957633887 0 -423 0 -1.3868607 0.011487602 0.016669034799060203 0 -428 0 -1.4530833 0.009660303 0.01400462463965267 0 -429 0 -1.4552101 0.009606653 0.013926471498510302 0 -430 0 -1.4554932 0.009599534 0.013916101323823175 0 -434 0 1.5582153 0.96550226 4.8573544264282766 1 -436 1 1.4286342 0.9520814 0.070843196999222466 1 -439 0 -1.4900287 0.008769167 0.012707030067382788 0 -440 1 2.5288706 0.99726295 0.0039541354596146936 1 -441 0 -1.2792237 0.01521197 0.022114868206894518 0 -442 0 -1.3768251 0.011792851 0.017114601690217213 0 -449 1 2.6800084 0.9981629 0.0026527760618351849 1 -450 0 -1.4375317 0.010061713 0.014589504767042908 0 -451 0 -1.4900287 0.008769167 0.012707030067382788 0 -452 0 -1.4860564 0.008860931 0.01284059512632533 0 -453 1 2.4471974 0.9966055 0.0049055365856787861 1 -454 0 -1.5048658 0.008434669 0.012220265005852901 0 -455 1 0.55157804 0.6615563 0.59606414877619884 1 -456 1 2.5915837 0.9976802 0.0033506703541261187 1 -457 1 2.656286 0.99804425 0.0028243100900196909 1 -464 0 -1.4914907 0.008735631 0.012658220654856148 0 -465 1 2.643771 0.99797857 0.0029192613497825144 1 -466 1 2.5306222 0.9972756 0.0039358553807143986 1 -467 1 2.1015139 0.9915761 0.012204631950163284 1 -474 0 -1.4900287 0.008769167 0.012707030067382788 0 -480 0 -1.4803663 0.008994038 0.013034358354277254 0 -482 1 4.264226 0.9999721 4.024450161416503E-05 1 -483 1 2.8133154 0.9987079 0.0018653253146604482 1 -484 0 -1.4612216 0.009456603 0.013707911824917898 0 -487 1 3.399762 0.99972564 0.0003958723792529149 1 -489 1 0.5020889 0.6316704 0.66275609042761874 1 -492 0 -1.4438386 0.009896969 0.014349434373091656 0 -493 1 2.5555058 0.9974486 0.0036855623160472976 1 -495 0 -1.4445758 0.0098778885 0.014321631501115688 0 -497 0 -1.4924375 0.008713981 0.01262671208698201 0 -501 0 -1.4531322 0.009659066 0.014002822915564145 0 -502 0 -1.4496942 0.009746407 0.014130064357101572 0 -504 0 -1.4831115 0.008929576 0.012940517754445463 0 -507 0 -1.3284451 0.013380531 0.019434338841309472 0 -510 0 -1.4831115 0.008929576 0.012940517754445463 0 -513 0 -1.4445758 0.0098778885 0.014321631501115688 0 -514 1 2.5598867 0.9974779 0.0036432331971241036 1 -517 0 -1.4396616 0.010005775 0.014507985306325046 0 -519 1 2.054533 0.9904725 0.013811181434465422 1 -520 0 -1.4974787 0.008599596 0.012460248317170512 0 -521 0 -1.5016329 0.008506458 0.012324719628844068 0 -522 1 1.6672657 0.97391796 0.038127844211451045 1 -523 1 1.9021587 0.98581284 0.020614318798665057 1 -527 0 -1.4115138 0.010770508 0.015622843432160721 0 -528 0 -1.42047 0.010521111 0.015259167981680956 0 -529 0 -1.4438386 0.009896969 0.014349434373091656 0 -531 0 -1.4232016 0.010446191 0.01514993576396167 0 -532 0 -1.4078557 0.010874045 0.015773849973586711 0 -533 0 -1.4552015 0.009606869 0.013926786240718135 0 -534 0 -1.4552101 0.009606653 0.013926471498510302 0 -535 0 -1.3623999 0.012245699 0.017775871466984346 0 -538 0 -1.4531322 0.009659066 0.014002822915564145 0 -539 0 -1.4431121 0.009915808 0.014376884942399512 0 -540 0 -1.3927023 0.011313532 0.016415007798749291 0 -541 0 -1.4908748 0.008749744 0.012678761369425539 0 -544 0 -1.3988788 0.011132318 0.016150603942727701 0 -546 1 2.7132857 0.99831736 0.0024295797436085639 1 -547 0 -1.4742839 0.009138514 0.013244700283846353 0 -548 0 -1.4784826 0.00903854 0.013099144413854295 0 -549 1 1.81378 0.9821451 0.025991956501029815 1 -557 0 -1.483142 0.0089288615 0.012939477917220827 0 -558 0 -1.4552101 0.009606653 0.013926471498510302 0 -559 0 -1.4490774 0.009762159 0.014153012753613161 0 -560 0 -1.4049767 0.010956224 0.0158937175406033 0 -561 0 -1.4049767 0.010956224 0.0158937175406033 0 -563 0 -1.4552015 0.009606869 0.013926786240718135 0 -565 1 3.0412924 0.99929243 0.0010211643377749521 1 -566 0 -1.3866773 0.011493112 0.01667707601868992 0 -569 1 2.880178 0.99891704 0.001563222996104044 1 -577 0 -1.4530833 0.009660303 0.01400462463965267 0 -578 0 -1.4530833 0.009660303 0.01400462463965267 0 -581 1 2.311439 0.9951468 0.0070187170806151992 1 -582 1 2.0334299 0.98993134 0.014599621900031615 1 -584 0 -1.4794143 0.009016501 0.013067059473830717 0 -586 1 3.442149 0.9997547 0.00035389770672077098 1 -590 1 1.7012169 0.9761037 0.034893634826877851 1 -593 0 -1.4612216 0.009456603 0.013707911824917898 0 -594 1 1.6729125 0.97429454 0.037570109159839701 1 -600 0 -1.4552015 0.009606869 0.013926786240718135 0 -602 0 -1.4531322 0.009659066 0.014002822915564145 0 -604 1 1.8324564 0.9829907 0.024750353718897031 1 -606 0 -1.4885982 0.008802104 0.012754969514873694 0 -607 0 -1.4831115 0.008929576 0.012940517754445463 0 -609 0 -1.4900287 0.008769167 0.012707030067382788 0 -612 1 4.263532 0.99997205 4.0330495340952001E-05 1 -613 0 -1.457992 0.009536925 0.013824902890591595 0 -614 0 -1.443208 0.009913321 0.014373260210539624 0 +323 1 1.18881 0.9277949 0.1081222199080385 1 +327 0 -1.2272866 0.008596861 0.012456267896939406 0 +328 1 1.220107 0.9338798 0.098691234085670126 1 +329 1 1.66271 0.98175305 0.026567918103568188 1 +331 0 -1.2315143 0.008488657 0.01229881767874451 0 +332 0 -1.2255177 0.0086425375 0.012522738424082254 0 +333 1 1.2068762 0.93136775 0.10257715968948981 1 +336 1 1.2241237 0.9346253 0.097539961291283228 1 +338 0 -1.2033671 0.009235303 0.013385631482353349 0 +343 0 -1.228457 0.00856677 0.012412479933523967 0 +344 1 2.1451018 0.9956924 0.0062280167287575711 1 +346 0 -1.2456083 0.008137599 0.011788101931998083 0 +347 0 -1.1820589 0.009843477 0.014271491879530612 0 +348 1 0.7704496 0.7839926 0.35108807901104988 1 +349 1 1.1994635 0.92992204 0.10481831527811435 1 +350 0 -1.2781533 0.0073810243 0.010688060280818633 0 +352 0 0.93649113 0.8570297 2.8062123778990693 1 +353 1 1.7637384 0.9864888 0.019625395371069238 1 +354 0 -1.2272866 0.008596861 0.012456267896939406 0 +355 0 -1.2675227 0.0076201446 0.011035645086907067 0 +358 1 1.6101708 0.97868043 0.031090240838711179 1 +360 1 2.9817219 0.9996548 0.0004980617286265759 1 +361 1 1.4490199 0.9657629 0.05025903200911358 1 +366 1 2.5738192 0.998817 0.0017076801051597557 1 +368 0 -1.1774392 0.009980475 0.014471116238714178 0 +370 0 -1.2461263 0.0081249755 0.01176974128592498 0 +371 0 -1.1774392 0.009980475 0.014471116238714178 0 +373 0 -1.2674762 0.007621207 0.011037189246739893 0 +376 0 -1.2272866 0.008596861 0.012456267896939406 0 +377 0 -1.2033671 0.009235303 0.013385631482353349 0 +378 0 -1.2658771 0.00765784 0.011090446509823042 0 +379 0 -1.1922754 0.0095470855 0.013839702946558922 0 +381 1 1.9371408 0.991955 0.011653446365992557 1 +383 0 -1.2414014 0.008240851 0.01193829304157505 0 +384 0 -1.2414014 0.008240851 0.01193829304157505 0 +387 0 -1.2081174 0.0091048805 0.013195730301881129 0 +388 0 -1.2273811 0.008594427 0.01245272659256396 0 +389 0 -1.2345154 0.008412668 0.012188253806122278 0 +391 1 1.898664 0.990972 0.013083825881113176 1 +392 0 -1.2484847 0.008067742 0.011686496825873889 0 +395 0 -1.2484847 0.008067742 0.011686496825873889 0 +396 0 -1.2406003 0.008260659 0.011967108123699813 0 +398 0 -1.2365414 0.008361751 0.012114175751818772 0 +399 0 -1.1776428 0.009974397 0.014462259442635747 0 +404 0 -1.2237978 0.008687181 0.012587708886769421 0 +406 0 -1.2274588 0.008592426 0.012449814137117746 0 +409 0 -1.2204521 0.008774682 0.012715057367079423 0 +413 0 -1.223793 0.008687305 0.012587889153513653 0 +414 1 1.4889055 0.96953195 0.044639648171584105 1 +415 0 -1.1248463 0.011679551 0.016949203122139926 0 +416 1 1.931327 0.99181354 0.011859173408493181 1 +418 0 -1.149593 0.010847138 0.015734605388814209 0 +419 0 -1.3097785 0.0067128236 0.0097172089442897278 0 +422 0 -1.162309 0.010442506 0.015144562966172998 0 +423 0 -1.2067921 0.009141082 0.013248438798178453 0 +428 0 -1.2272866 0.008596861 0.012456267896939406 0 +429 0 -1.2384154 0.008314926 0.012046053477396252 0 +430 0 -1.2160814 0.008890305 0.012883352222030494 0 +434 0 1.3154399 0.9495954 4.3103005513864838 1 +436 1 1.3833226 0.95855594 0.061065471644374987 1 +439 0 -1.2737958 0.007478125 0.010829195442343941 0 +440 1 1.9225917 0.99159646 0.01217497337758197 1 +441 0 -1.1378857 0.011233363 0.016298030731159381 0 +442 0 -1.1551075 0.010669793 0.01547596774766722 0 +449 1 2.0103436 0.9935412 0.0093483281981078716 1 +450 0 -1.2402703 0.008268834 0.011979000660228434 0 +451 0 -1.2737958 0.007478125 0.010829195442343941 0 +452 0 -1.2581867 0.007836475 0.011350174667777871 0 +453 1 1.872927 0.990249 0.014136787508837099 1 +454 0 -1.2647548 0.007683656 0.011127979408058476 0 +455 1 0.65830463 0.7211563 0.47161612041525247 1 +456 1 1.9675331 0.99265575 0.010634606714884054 1 +457 1 2.1939058 0.9962808 0.0053756895933591686 1 +464 0 -1.264048 0.007699958 0.011151680338550251 0 +465 1 2.1614888 0.9958996 0.0059277626193753588 1 +466 1 1.9563407 0.992405 0.010999094812038849 1 +467 1 1.6580784 0.9815006 0.026938907619080441 1 +474 0 -1.2737958 0.007478125 0.010829195442343941 0 +480 0 -1.2516133 0.007992436 0.011576974026374717 0 +482 1 3.3211005 0.9998762 0.00017861503931714426 1 +483 1 2.1265 0.9954444 0.0065873329922731634 1 +484 0 -1.2456272 0.0081371395 0.011787434095638438 0 +487 1 2.5093064 0.9985628 0.0020749142162757184 1 +489 1 0.60993165 0.69083583 0.53358517770758007 1 +492 0 -1.235756 0.008381452 0.012142838525431235 0 +493 1 1.9121976 0.9913306 0.012561796793150075 1 +495 0 -1.2260168 0.008629627 0.012503949635303229 0 +497 0 -1.2633969 0.0077150064 0.011173559091089009 0 +501 0 -1.2575177 0.007852209 0.011373053157156589 0 +502 0 -1.2595898 0.007803578 0.011302340106673026 0 +504 0 -1.228457 0.00856677 0.012412479933523967 0 +507 0 -1.1100978 0.012205357 0.017716949770581924 0 +510 0 -1.228457 0.00856677 0.012412479933523967 0 +513 0 -1.2260168 0.008629627 0.012503949635303229 0 +514 1 1.9611464 0.9925137 0.010841054911793607 1 +517 0 -1.2033671 0.009235303 0.013385631482353349 0 +519 1 1.6117868 0.97878206 0.030940439541602423 1 +520 0 -1.2530813 0.00795734 0.011525934222332975 0 +521 0 -1.2864741 0.0071990686 0.010423625881359583 0 +522 1 1.4433907 0.96519595 0.051106227177230769 1 +523 1 1.5021174 0.9706895 0.04291824492023022 1 +527 0 -1.2479138 0.008081559 0.011706592959255056 0 +528 0 -1.2269958 0.008604352 0.012467169693719775 0 +529 0 -1.235756 0.008381452 0.012142838525431235 0 +531 0 -1.2274588 0.008592426 0.012449814137117746 0 +532 0 -1.2001765 0.0093239425 0.013514709124084631 0 +533 0 -1.2484847 0.008067742 0.011686496825873889 0 +534 0 -1.2384154 0.008314926 0.012046053477396252 0 +535 0 -1.1694435 0.010222062 0.014823209102968853 0 +538 0 -1.2575177 0.007852209 0.011373053157156589 0 +539 0 -1.2725711 0.0075056446 0.010869197597557794 0 +540 0 -1.2250142 0.0086555835 0.012541723982681326 0 +541 0 -1.2532576 0.007953135 0.011519819156242013 0 +544 0 -1.1860404 0.009726902 0.014101647005927349 0 +546 1 2.0275328 0.9938661 0.0088766194277279178 1 +547 0 -1.2170439 0.008864714 0.01284610169197423 0 +548 0 -1.2296379 0.008536515 0.012368454786871061 0 +549 1 1.5769174 0.97647977 0.034337941018001673 1 +557 0 -1.2781533 0.0073810243 0.010688060280818633 0 +558 0 -1.2384154 0.008314926 0.012046053477396252 0 +559 0 -1.2655382 0.0076656267 0.011101767206480718 0 +560 0 -1.2542456 0.007929614 0.011485613187028508 0 +561 0 -1.2542456 0.007929614 0.011485613187028508 0 +563 0 -1.2484847 0.008067742 0.011686496825873889 0 +565 1 2.2523775 0.9968813 0.0045063547491217505 1 +566 0 -1.2113973 0.0090158945 0.013066176822659941 0 +569 1 2.27636 0.9970987 0.0041917975654111308 1 +577 0 -1.2272866 0.008596861 0.012456267896939406 0 +578 0 -1.2272866 0.008596861 0.012456267896939406 0 +581 1 1.7694135 0.9867155 0.019293929430500921 1 +582 1 1.6290632 0.97983956 0.029382549822020751 1 +584 0 -1.3060807 0.006787741 0.0098260265281388329 0 +586 1 2.509679 0.9985644 0.0020725891107481044 1 +590 1 1.5352468 0.97340745 0.038884282468759225 1 +593 0 -1.2456272 0.0081371395 0.011787434095638438 0 +594 1 1.35704 0.95528334 0.065999384718859078 1 +600 0 -1.2484847 0.008067742 0.011686496825873889 0 +602 0 -1.2575177 0.007852209 0.011373053157156589 0 +604 1 1.441893 0.9650436 0.051333964534827747 1 +606 0 -1.2663386 0.0076472503 0.011075051127751125 0 +607 0 -1.228457 0.00856677 0.012412479933523967 0 +609 0 -1.2737958 0.007478125 0.010829195442343941 0 +612 1 3.1690764 0.999804 0.0002827671872904593 1 +613 0 -1.2373637 0.008341172 0.012084235861254811 0 +614 0 -1.2152277 0.0089130625 0.012916479675079178 0 617 0 ? ? ? 0 -618 0 -1.4531322 0.009659066 0.014002822915564145 0 -619 0 -1.4490774 0.009762159 0.014153012753613161 0 -621 0 0.75065064 0.76791567 2.1072789533873553 1 -622 0 -1.3915341 0.011348133 0.016465499216567731 0 -624 0 -1.437456 0.010063706 0.014592409329757701 0 -627 0 -1.278245 0.0152507825 0.02217172941394379 0 -629 0 -1.4914907 0.008735631 0.012658220654856148 0 -633 1 1.4559345 0.95526904 0.066020988856273249 1 -634 0 -1.4908748 0.008749744 0.012678761369425539 0 -638 0 -1.4914907 0.008735631 0.012658220654856148 0 -639 0 -1.483142 0.0089288615 0.012939477917220827 0 -641 0 -1.4552015 0.009606869 0.013926786240718135 0 -642 0 -1.4552015 0.009606869 0.013926786240718135 0 -644 0 -1.4881064 0.008813457 0.012771493716677452 0 -645 0 -1.4552015 0.009606869 0.013926786240718135 0 -649 0 -1.4552015 0.009606869 0.013926786240718135 0 -652 0 -1.4452621 0.009860157 0.014295795448825529 0 -653 0 -1.4531322 0.009659066 0.014002822915564145 0 -654 0 -1.416158 0.010640464 0.015433198933891867 0 -656 0 -1.4490774 0.009762159 0.014153012753613161 0 -657 0 0.77659476 0.7799156 2.1838710182908652 1 -660 0 -1.4530833 0.009660303 0.01400462463965267 0 -661 0 -1.4115138 0.010770508 0.015622843432160721 0 -665 0 -1.4831115 0.008929576 0.012940517754445463 0 -668 1 1.4594649 0.9556662 0.065421324144804505 1 -670 1 1.923006 0.98656344 0.019516263807251893 1 -678 0 -1.4831115 0.008929576 0.012940517754445463 0 -679 0 -1.4881064 0.008813457 0.012771493716677452 0 -680 1 4.0927296 0.9999561 6.3376998885288649E-05 1 -681 1 2.4021688 0.996178 0.005524586061806897 1 -682 0 -1.400563 0.011083404 0.016079243303635436 0 -683 0 -1.4831115 0.008929576 0.012940517754445463 0 -685 0 -1.4831115 0.008929576 0.012940517754445463 0 -688 0 -1.4914907 0.008735631 0.012658220654856148 0 -689 0 -1.2336183 0.017127996 0.024924543698761008 0 -691 1 1.4977734 0.95976484 0.059247135195144636 1 -692 0 -1.4908748 0.008749744 0.012678761369425539 0 -693 0 -1.4178497 0.01059348 0.015364688902515364 0 -694 0 -1.4480503 0.009788445 0.014191310645197746 0 -696 1 2.2876167 0.99483293 0.0074738275658872465 1 -697 1 1.7867663 0.9808485 0.027897790132007501 1 -698 1 2.0381904 0.990056 0.014417996634455362 1 -0 0 -1.1330655 0.007665574 0.011101690705840397 0 -1 0 0.9337132 0.91784257 3.6054650836837365 1 -2 0 -1.1207929 0.00800135 0.011589937427300587 0 -3 0 0.9170699 0.9133143 3.5280618844247424 1 -4 0 -1.109437 0.008325023 0.012060741808292988 0 -7 0 -1.123038 0.007938851 0.011499045732174347 0 -12 1 0.7800009 0.86671245 0.20637466470729302 1 -13 0 -1.1074398 0.008383277 0.012145492916229685 0 -14 1 1.2002811 0.9661676 0.049654578489145548 1 -15 1 0.9246728 0.9154102 0.12750969506171719 1 -16 0 -1.1276242 0.007812686 0.011315584066679825 0 -17 0 -1.1306075 0.007731685 0.0111978084971893 0 -19 0 -1.1357371 0.0075943535 0.010998151145919928 0 -22 0 -1.1253841 0.00787406 0.011404828234943619 0 +618 0 -1.2575177 0.007852209 0.011373053157156589 0 +619 0 -1.2655382 0.0076656267 0.011101767206480718 0 +621 0 0.80149895 0.7994584 2.3180264456106543 1 +622 0 -1.251737 0.007989473 0.011572664205581805 0 +624 0 -1.2465595 0.008114432 0.011754405706261203 0 +627 0 -1.0857006 0.013126925 0.01906354750547521 0 +629 0 -1.264048 0.007699958 0.011151680338550251 0 +633 1 1.3667102 0.9565151 0.064140390622424909 1 +634 0 -1.2532576 0.007953135 0.011519819156242013 0 +638 0 -1.264048 0.007699958 0.011151680338550251 0 +639 0 -1.2781533 0.0073810243 0.010688060280818633 0 +641 0 -1.2484847 0.008067742 0.011686496825873889 0 +642 0 -1.2484847 0.008067742 0.011686496825873889 0 +644 0 -1.2414014 0.008240851 0.01193829304157505 0 +645 0 -1.2484847 0.008067742 0.011686496825873889 0 +649 0 -1.2484847 0.008067742 0.011686496825873889 0 +652 0 -1.2542933 0.00792848 0.011483964939196149 0 +653 0 -1.2575177 0.007852209 0.011373053157156589 0 +654 0 -1.2406003 0.008260659 0.011967108123699813 0 +656 0 -1.2655382 0.0076656267 0.011101767206480718 0 +657 0 0.85161436 0.82264227 2.4952658580582621 1 +660 0 -1.2272866 0.008596861 0.012456267896939406 0 +661 0 -1.2479138 0.008081559 0.011706592959255056 0 +665 0 -1.228457 0.00856677 0.012412479933523967 0 +668 1 1.3882962 0.9591489 0.060173320022865261 1 +670 1 1.5648978 0.97563094 0.035592585580711114 1 +678 0 -1.228457 0.00856677 0.012412479933523967 0 +679 0 -1.2414014 0.008240851 0.01193829304157505 0 +680 1 3.022023 0.9996944 0.00044094490375990293 1 +681 1 1.8298862 0.98890954 0.016089533931025392 1 +682 0 -1.2428946 0.008204054 0.011884766716602314 0 +683 0 -1.228457 0.00856677 0.012412479933523967 0 +685 0 -1.228457 0.00856677 0.012412479933523967 0 +688 0 -1.264048 0.007699958 0.011151680338550251 0 +689 0 -1.1009516 0.012543106 0.018210324757234331 0 +691 1 1.3482523 0.9541351 0.067734506824647087 1 +692 0 -1.2532576 0.007953135 0.011519819156242013 0 +693 0 -1.2301259 0.008524043 0.012350306290290665 0 +694 0 -1.235653 0.00838404 0.012146602640017302 0 +696 1 1.9361792 0.99193174 0.011687255369906764 1 +697 1 1.5745599 0.97631556 0.034580573821226458 1 +698 1 1.7136495 0.9843155 0.022807260688935727 1 +0 0 -1.1396108 0.011449234 0.01661303838020094 0 +1 0 0.98902833 0.92953086 3.8268645615308938 1 +2 0 -1.1361089 0.011581021 0.016805382758276803 0 +3 0 0.97100323 0.9255258 3.7471151682034405 1 +4 0 -1.0785714 0.0139737055 0.020301975179298343 0 +7 0 -1.1245402 0.012027132 0.017456672721167559 0 +12 1 -1.052682 0.0152035095 6.0394518052746244 0 +13 0 -1.1321943 0.011730115 0.017023015948818703 0 +14 1 1.2364503 0.9676277 0.047476019184346567 1 +15 1 0.9731152 0.92600566 0.11090708091039683 1 +16 0 -1.1383487 0.011496558 0.016682105201821239 0 +17 0 -1.1373855 0.011532806 0.016735009192845141 0 +19 0 -1.1421459 0.011354756 0.016475163360988903 0 +22 0 -1.1366465 0.0115606915 0.016775709655013345 0 23 1 ? ? ? 0 -24 0 -1.124339 0.007902858 0.011446704476681715 0 -26 0 -1.1032244 0.008507558 0.012326320052336947 0 -27 0 -1.1300513 0.007746724 0.011219674408243712 0 -29 0 -1.1118658 0.008254722 0.011958471260980198 0 -30 0 -1.1144156 0.00818155 0.011852031423018961 0 -33 0 -1.1154537 0.008151945 0.011808968926040345 0 -34 0 -1.1222382 0.007961062 0.011531346382182695 0 -36 1 1.309447 0.9767124 0.033994273713638244 1 -38 1 1.0344237 0.94092023 0.087855670854846829 1 -39 1 0.83703864 0.88825274 0.17095786892853884 1 -42 1 1.1177064 0.95526326 0.066029720620280202 1 -43 1 -1.0914044 0.008865871 6.8175219890817687 0 -47 0 -1.1213936 0.0079845805 0.011565549372462803 0 -49 1 1.0305241 0.94015235 0.089033537480907365 1 -53 1 1.1563569 0.9607315 0.057794794567089018 1 -55 1 1.0800781 0.94924754 0.075143741274176931 1 -57 1 0.9589613 0.9243016 0.11356437650305147 1 -58 1 0.88427603 0.90372616 0.1460424093039884 1 -59 1 0.9442126 0.92058736 0.11937345957228779 1 -61 0 -1.1127788 0.008228447 0.011920250207795814 0 -62 1 1.1015996 0.9527762 0.069790728890640016 1 -65 1 1.1100354 0.9540947 0.067795612796055102 1 -67 1 0.9810804 0.92957366 0.10535891271341713 1 -75 0 -1.1174511 0.008095282 0.011726552559967277 0 -78 0 -1.1132877 0.008213838 0.011898998212079182 0 -80 0 -1.1127793 0.008228433 0.011920229886365239 0 -81 0 -1.1169924 0.00810826 0.011745428758102684 0 -83 0 -1.1332204 0.0076614246 0.011095657999977108 0 -84 1 1.1783425 0.96354985 0.053568783936532305 1 -85 1 1.0753026 0.9484314 0.07638470915813507 1 -86 1 0.87544775 0.9009877 0.15042070838808383 1 -87 1 1.0354706 0.9411248 0.087542052184469282 1 -89 0 -1.1234512 0.007927401 0.011482395245599426 0 -94 0 -1.1262618 0.007849955 0.011369775879992286 0 -101 1 0.68830734 0.8248218 0.27784557920190756 1 -103 1 -1.1696934 0.006744369 7.2121008273478227 0 -107 1 1.0195283 0.9379366 0.092437681904234417 1 -110 0 -1.0996349 0.008614826 0.012482411205087062 0 -114 0 -1.0880932 0.008968905 0.012997769610086835 0 -116 0 -1.1128553 0.008226248 0.011917051618146355 0 -118 0 -1.1226587 0.007949376 0.011514351510015829 0 -119 0 -1.109485 0.008323626 0.012058709468750045 0 -124 1 1.1565694 0.9607597 0.057752458805352014 1 -126 1 1.1509618 0.96000856 0.058880822658874232 1 -127 0 -1.1283454 0.007793028 0.01128700001539292 0 -130 0 -1.1241809 0.007907222 0.011453050822155073 0 -134 0 -1.1362723 0.007580164 0.010977523327413559 0 -135 0 -1.1299217 0.007750232 0.011224774651226633 0 -136 0 -1.1276242 0.007812686 0.011315584066679825 0 +24 0 -1.1323191 0.011725332 0.017016033253838249 0 +26 0 -1.1097994 0.012620236 0.018323016729344839 0 +27 0 -1.1403185 0.011422779 0.016574431502966412 0 +29 0 -1.1169518 0.012328933 0.01789744681324926 0 +30 0 -1.1279317 0.011894619 0.017263182221643397 0 +33 0 -1.1043559 0.01284648 0.0186536280971712 0 +34 0 -1.1273962 0.011915445 0.017293590135388604 0 +36 1 1.3386434 0.97669303 0.034022887518223757 1 +38 1 1.0785991 0.94662946 0.079128267033883398 1 +39 1 0.9278369 0.9150713 0.12804392244853399 1 +42 1 1.1991884 0.9635382 0.053586186678646036 1 +43 1 -0.92657876 0.022888208 5.4492516883288653 0 +47 0 -1.133943 0.01166328 0.016925452979744748 0 +49 1 1.09097 0.9486585 0.076039219001838895 1 +53 1 1.2137656 0.9651941 0.051108989034685061 1 +55 1 1.1261102 0.95403117 0.067891693200425082 1 +57 1 1.0715668 0.9454425 0.080938379453869461 1 +58 1 0.9660017 0.9243779 0.11344529810781659 1 +59 1 1.0359164 0.9390336 0.090751272073977901 1 +61 0 -1.1337178 0.011671865 0.016937984631799692 0 +62 1 1.1707509 0.9600871 0.058762769618043192 1 +65 1 1.1659532 0.95947486 0.059683086787269253 1 +67 1 1.08381 0.9474932 0.077812512347655127 1 +75 0 -1.1272507 0.011921112 0.017301864647317648 0 +78 0 -1.1367682 0.011556094 0.016768998662134882 0 +80 0 -1.0945407 0.013264586 0.019264806960499054 0 +81 0 -1.1088833 0.012658035 0.01837824808604005 0 +83 0 -1.1343621 0.011647317 0.016902150486775704 0 +84 1 1.2403295 0.96802706 0.04688072512608029 1 +85 1 1.1400868 0.9560157 0.064893774932294118 1 +86 1 0.93784416 0.9176076 0.12405074591585261 1 +87 1 1.1039943 0.9507158 0.072913987973693792 1 +89 0 -1.1143463 0.012434275 0.018051328345130388 0 +94 0 -1.1337652 0.011670057 0.016935344519875207 0 +101 1 0.7253401 0.8465357 0.24035721384507364 1 +103 1 -1.1579578 0.010782635 6.5351464323445754 0 +107 1 1.0749816 0.9460219 0.080054491081168397 1 +110 0 -1.082407 0.013800049 0.020047914023123871 0 +114 0 -1.0914135 0.0134005975 0.019463681270144525 0 +116 0 -1.1049641 0.012821004 0.018616395692546056 0 +118 0 -1.152132 0.010990033 0.015943034595105907 0 +119 0 -1.1178445 0.012293047 0.017845029350003233 0 +124 1 1.2380968 0.9677978 0.047222411653718974 1 +126 1 1.2041231 0.9641071 0.052734680208221141 1 +127 0 -1.1354455 0.011606155 0.016842068197582008 0 +130 0 -1.1220528 0.012125248 0.017599954141935562 0 +134 0 -1.1452729 0.011239281 0.016306665075714037 0 +135 0 -1.1528184 0.010965393 0.01590709202006884 0 +136 0 -1.1383487 0.011496558 0.016682105201821239 0 139 0 ? ? ? 0 -140 0 -1.1223897 0.007956849 0.011525220458022329 0 -142 1 1.0598775 0.9457096 0.080530874277415593 1 -143 0 -1.0830694 0.009127492 0.013228652038033533 0 -146 1 0.7958858 0.87304187 0.19587725247558344 1 -148 0 -1.1822221 0.006455211 0.0093430892074493916 0 -149 1 1.2765069 0.97392344 0.03811972116620118 1 -153 0 -1.109682 0.008317905 0.012050386390746052 0 -155 1 0.922753 0.91488534 0.12833714571823487 1 -157 0 -1.1262618 0.007849955 0.011369775879992286 0 +140 0 -1.138028 0.011508613 0.016699699325297319 0 +142 1 1.0999073 0.9500788 0.073880940790039151 1 +143 0 -1.1335585 0.011677941 0.01694685392062556 0 +146 1 -1.0611465 0.014790134 6.0792211079665321 0 +148 0 -1.1642773 0.010562035 0.015318838105258355 0 +149 1 1.3456196 0.97721237 0.033255969698413552 1 +153 0 -1.1406034 0.011412147 0.016558915648592666 0 +155 1 0.9575066 0.922391 0.11654966261709278 1 +157 0 -1.1337652 0.011670057 0.016935344519875207 0 158 0 ? ? ? 0 -159 1 1.2999746 0.97594166 0.035133189108404989 1 -160 1 1.2043302 0.9666306 0.048963461708256155 1 -162 0 -1.1283454 0.007793028 0.01128700001539292 0 -163 0 -1.1283889 0.0077918433 0.011285277515329178 0 -165 0 -1.1251904 0.00787939 0.0114125787894873 0 -166 1 1.1199185 0.9555949 0.065528944842205422 1 -168 0 -1.1283454 0.007793028 0.01128700001539292 0 -170 0 -1.1223897 0.007956849 0.011525220458022329 0 -172 0 -1.1213936 0.0079845805 0.011565549372462803 0 -175 1 1.1031612 0.95302296 0.069417128042735046 1 -178 0 -1.1306075 0.007731685 0.0111978084971893 0 -182 0 -1.1357371 0.0075943535 0.010998151145919928 0 -184 1 1.0567007 0.94513243 0.081411597279103534 1 -185 0 -1.1152303 0.008158307 0.011818222621464713 0 -186 1 0.9926462 0.93219334 0.10129889097007028 1 -190 1 1.3113946 0.97686785 0.033764679484254864 1 -193 0 -1.124339 0.007902858 0.011446704476681715 0 -194 0 -1.1283454 0.007793028 0.01128700001539292 0 -195 0 -1.1306075 0.007731685 0.0111978084971893 0 -197 0 -1.101768 0.00855092 0.012389415554594085 0 -200 1 1.2397302 0.97042525 0.043311008463096642 1 -203 0 -1.1330655 0.007665574 0.011101690705840397 0 -208 0 -1.1131408 0.008218052 0.011905128433067014 0 -213 1 1.3674624 0.9809328 0.027773829590120081 1 -214 1 1.3293985 0.9782578 0.031713423007557931 1 -215 1 1.1887612 0.9648165 0.051673498938228972 1 -217 0 -1.124339 0.007902858 0.011446704476681715 0 -220 0 -1.1134541 0.008209066 0.011892057885038583 0 -221 1 1.2833158 0.9745254 0.037228318316142726 1 -222 1 -1.0701756 0.00954735 6.7106839365311677 0 -224 1 1.2488513 0.9713331 0.04196198994754427 1 -225 0 -1.1213936 0.0079845805 0.011565549372462803 0 -227 1 1.1535391 0.96035546 0.058359599064123871 1 -229 1 1.328885 0.9782193 0.031770209154217889 1 -230 1 1.098734 0.95232016 0.070481423240788557 1 -231 1 1.1551619 0.9605724 0.058033706115406078 1 -232 0 0.930905 0.9170939 3.5923774652397524 1 -234 0 -1.1202629 0.008016176 0.011611499108274534 0 +159 1 1.361773 0.97837186 0.031545187390341137 1 +160 1 1.2648827 0.97044635 0.043279640154898026 1 +162 0 -1.1354455 0.011606155 0.016842068197582008 0 +163 0 -1.1377423 0.011519365 0.016715392039980122 0 +165 0 -1.1440386 0.011284723 0.016372970791161501 0 +166 1 1.1740303 0.96050054 0.058141672382886173 1 +168 0 -1.1354455 0.011606155 0.016842068197582008 0 +170 0 -1.138028 0.011508613 0.016699699325297319 0 +172 0 -1.133943 0.01166328 0.016925452979744748 0 +175 1 1.1718655 0.96022815 0.058550871624026037 1 +178 0 -1.1373855 0.011532806 0.016735009192845141 0 +182 0 -1.1421459 0.011354756 0.016475163360988903 0 +184 1 1.1118553 0.95191944 0.071088615414448192 1 +185 0 -1.124998 0.012009162 0.017430432319144644 0 +186 1 1.0722877 0.9455653 0.08075102736318554 1 +190 1 1.3522085 0.9776925 0.0325473314864273 1 +193 0 -1.1323191 0.011725332 0.017016033253838249 0 +194 0 -1.1354455 0.011606155 0.016842068197582008 0 +195 0 -1.1373855 0.011532806 0.016735009192845141 0 +197 0 -1.0514342 0.015265406 0.02219315374262016 0 +200 1 1.3015174 0.97372913 0.03840758741601602 1 +203 0 -1.1396108 0.011449234 0.01661303838020094 0 +208 0 -1.1234338 0.012070676 0.017520259567070942 0 +213 1 1.3996947 0.9808717 0.027863686770957376 1 +214 1 1.3684912 0.9788369 0.030859614803626791 1 +215 1 1.2307085 0.9670277 0.048370843861697058 1 +217 0 -1.1323191 0.011725332 0.017016033253838249 0 +220 0 -1.1029195 0.012906848 0.01874185728410745 0 +221 1 1.3759619 0.9793425 0.03011457129583596 1 +222 1 -1.0770677 0.014042373 6.1540694535878515 0 +224 1 1.3045129 0.9739813 0.038033990509864411 1 +225 0 -1.133943 0.01166328 0.016925452979744748 0 +227 1 1.2025375 0.96392524 0.05300683283700653 1 +229 1 1.4105057 0.9815309 0.026894401361644876 1 +230 1 1.1660347 0.95948535 0.059667313168651898 1 +231 1 1.2273237 0.966669 0.048906083736544328 1 +232 0 1.0201598 0.9359821 3.9653810202137993 1 +234 0 -1.1255153 0.011988883 0.017400819792906051 0 235 0 ? ? ? 0 -236 1 1.2969778 0.97569263 0.035501364398050017 1 -238 1 1.2898574 0.975091 0.036391259578451532 1 -243 0 -1.0980439 0.0086627975 0.012552222529201768 0 -245 0 -1.1162674 0.008128813 0.011775322334086517 0 -251 1 1.202626 0.9664365 0.049253144108785021 1 -253 1 1.1177064 0.95526326 0.066029720620280202 1 -255 1 0.93955636 0.9193806 0.12126586276235005 1 -256 0 -1.1223897 0.007956849 0.011525220458022329 0 -261 1 1.2494476 0.9713915 0.041875233955906838 1 -263 1 1.2238339 0.9687759 0.045765077250808577 1 -264 1 1.0892453 0.95078015 0.072816306322249127 1 -265 0 -1.125992 0.007857359 0.011380542169694833 0 -266 1 1.1041114 0.9531725 0.06919075856459983 1 -270 1 1.08556 0.9501693 0.073743463138861348 1 -273 1 0.84026736 0.88937616 0.16913435591210796 1 -274 0 -1.1297894 0.007753815 0.011229984595470687 0 -281 0 -1.1154537 0.008151945 0.011808968926040345 0 -282 1 0.91142255 0.9117271 0.13332608247235195 1 -286 1 1.4042004 0.98320735 0.02443240154429947 1 -287 0 -1.1362723 0.007580164 0.010977523327413559 0 -289 1 1.1239699 0.95619625 0.064621349352190452 1 +236 1 1.3499153 0.9775265 0.032792302352076044 1 +238 1 1.3130698 0.9746887 0.036986563208653886 1 +243 0 -1.1311435 0.011770459 0.017081912029459148 0 +245 0 -1.1120358 0.012528431 0.018188883295371855 0 +251 1 1.2312171 0.9670813 0.048290904002777607 1 +253 1 1.1991884 0.9635382 0.053586186678646036 1 +255 1 1.0076021 0.9334488 0.099357215735069879 1 +256 0 -1.138028 0.011508613 0.016699699325297319 0 +261 1 1.3242589 0.9755855 0.035659749210537134 1 +263 1 1.2661349 0.97056484 0.04310349407597356 1 +264 1 1.1220188 0.9534342 0.06879467487765234 1 +265 0 -1.1182946 0.01227499 0.01781865401412458 0 +266 1 1.1533208 0.95781934 0.062174524618938971 1 +270 1 1.1568015 0.9582818 0.061478103446905089 1 +273 1 0.8922878 0.9054802 0.14324499098348753 1 +274 0 -1.140697 0.011408658 0.016553823016225946 0 +281 0 -1.1043559 0.01284648 0.0186536280971712 0 +282 1 0.9580284 0.9225144 0.11635669657538454 1 +286 1 1.4260923 0.9824423 0.025555387738256136 1 +287 0 -1.1452729 0.011239281 0.016306665075714037 0 +289 1 1.1751524 0.9606411 0.057930581703106672 1 292 1 ? ? ? 0 294 0 ? ? ? 0 -295 1 1.1071836 0.953653 0.068463711338507066 1 -298 0 -1.1546983 0.007107404 0.010290429184217832 0 -302 1 1.4213269 0.98417443 0.023014060289988692 1 -305 1 1.1708379 0.9626104 0.054976049369756919 1 -306 0 -1.124339 0.007902858 0.011446704476681715 0 -307 0 -1.124339 0.007902858 0.011446704476681715 0 -310 0 -1.134364 0.0076308744 0.011051243817366939 0 -313 0 -1.1184047 0.008068365 0.011687403015130963 0 +295 1 1.169635 0.95994556 0.058975504953637542 1 +298 0 -1.1542333 0.010914776 0.015833258832872014 0 +302 1 1.4559847 0.9840683 0.023169681901452584 1 +305 1 1.2200643 0.965887 0.050073663079026748 1 +306 0 -1.1323191 0.011725332 0.017016033253838249 0 +307 0 -1.1323191 0.011725332 0.017016033253838249 0 +310 0 -1.1439253 0.011288901 0.016379067052989027 0 +313 0 -1.1357223 0.011595662 0.016826752011794854 0 315 0 ? ? ? 0 -318 0 -1.1574371 0.007039678 0.010192024801420995 0 -320 1 1.0976412 0.95214516 0.070746558583319907 1 -322 0 -1.1283454 0.007793028 0.01128700001539292 0 -324 0 -1.124339 0.007902858 0.011446704476681715 0 -325 0 -1.113921 0.008195692 0.011872602658637672 0 -326 1 1.0526656 0.944391 0.082543783443699961 1 -330 1 1.111193 0.9542729 0.067526153118338039 1 -334 1 1.0557685 0.94496197 0.081671833160943022 1 -335 0 -1.1184047 0.008068365 0.011687403015130963 0 -337 0 -1.124339 0.007902858 0.011446704476681715 0 -339 1 1.0469 0.9433153 0.084187985813185512 1 -340 1 1.1337848 0.95762104 0.062473247715344418 1 -341 0 -1.124339 0.007902858 0.011446704476681715 0 -342 0 -1.1203263 0.0080144 0.011608916132203563 0 -345 0 -1.1184047 0.008068365 0.011687403015130963 0 -351 0 -1.1262618 0.007849955 0.011369775879992286 0 -356 1 -1.1234308 0.007927965 6.9788336234239656 0 -357 1 1.3021178 0.9761182 0.034872227535259197 1 -359 1 1.1163782 0.9550629 0.066332304386791108 1 -362 0 -1.1166741 0.008117278 0.011758545389116545 0 -363 0 -1.0625968 0.0098030055 0.014212524518944404 0 -364 0 -1.1262618 0.007849955 0.011369775879992286 0 -365 0 -1.1233132 0.007931225 0.011487956221288443 0 -367 1 1.2174531 0.96808916 0.046788165639959571 1 -369 0 -1.1037275 0.00849263 0.012304598621958016 0 -372 0 -1.124566 0.007896595 0.01143759672065953 0 -374 0 -1.1222382 0.007961062 0.011531346382182695 0 -375 0 -1.1184047 0.008068365 0.011687403015130963 0 -380 0 -1.1184047 0.008068365 0.011687403015130963 0 -382 0 -1.1047949 0.008461044 0.012258640356822125 0 -385 0 -1.1135433 0.008206511 0.011888340495412174 0 -386 1 1.0381248 0.94164044 0.086751819858459062 1 -390 0 -1.1105151 0.008293743 0.012015237056632821 0 -393 0 -1.1069949 0.008396308 0.012164451831112489 0 -394 0 -1.1035254 0.008498623 0.012313318864934643 0 -397 0 -1.1246129 0.007895299 0.011435712878225906 0 -400 1 1.1346995 0.9577515 0.062276695862690384 1 -401 0 -1.1223897 0.007956849 0.011525220458022329 0 -402 0 -1.1019315 0.008546042 0.012382318375568234 0 -403 0 -1.0986285 0.00864514 0.012526525215295091 0 -405 0 -1.1213936 0.0079845805 0.011565549372462803 0 -407 0 -1.1213936 0.0079845805 0.011565549372462803 0 -408 0 -1.082147 0.009156912 0.013271487081826362 0 -410 0 -1.1213936 0.0079845805 0.011565549372462803 0 +318 0 -1.1584249 0.0107661765 0.015616526255790365 0 +320 1 1.1710309 0.9601226 0.058709478731654927 1 +322 0 -1.1354455 0.011606155 0.016842068197582008 0 +324 0 -1.1323191 0.011725332 0.017016033253838249 0 +325 0 -1.1195756 0.012223745 0.017743806132878466 0 +326 1 1.1251012 0.9538846 0.068113351486251944 1 +330 1 1.1686294 0.95981747 0.05916802388005836 1 +334 1 1.1078148 0.95130426 0.072021260908277915 1 +335 0 -1.1357223 0.011595662 0.016826752011794854 0 +337 0 -1.1323191 0.011725332 0.017016033253838249 0 +339 1 1.093744 0.9491034 0.075362801954024769 1 +340 1 1.2145998 0.9652867 0.050970635387237055 1 +341 0 -1.1323191 0.011725332 0.017016033253838249 0 +342 0 -1.136776 0.011555801 0.016768571834800274 0 +345 0 -1.1357223 0.011595662 0.016826752011794854 0 +351 0 -1.1337652 0.011670057 0.016935344519875207 0 +356 1 -1.1241 0.01204444 6.375488869170824 0 +357 1 1.352685 0.9777268 0.032496671248591603 1 +359 1 1.157468 0.95836985 0.061345571083311158 1 +362 0 -1.1267911 0.011939023 0.017328015680589677 0 +363 0 -1.0302569 0.016354453 0.023789555183833525 0 +364 0 -1.1337652 0.011670057 0.016935344519875207 0 +365 0 -1.1351866 0.0116159795 0.016856408492649289 0 +367 1 1.2659607 0.9705484 0.043127947678585951 1 +369 0 -1.1064802 0.012757716 0.018523907507216861 0 +372 0 -1.1291512 0.011847323 0.017194128368617327 0 +374 0 -1.1273962 0.011915445 0.017293590135388604 0 +375 0 -1.1357223 0.011595662 0.016826752011794854 0 +380 0 -1.1357223 0.011595662 0.016826752011794854 0 +382 0 -1.084039 0.013726808 0.019940774892947058 0 +385 0 -1.0890983 0.013502176 0.019612226089607259 0 +386 1 1.1033953 0.9506229 0.073054914453943423 1 +390 0 -1.1044893 0.012840888 0.018645456077108418 0 +393 0 -1.1267003 0.011942566 0.017333189918116637 0 +394 0 -1.0971048 0.013154085 0.019103253085330004 0 +397 0 -1.1395046 0.011453209 0.016618840714141502 0 +400 1 1.1884639 0.9622718 0.055483630266295568 1 +401 0 -1.138028 0.011508613 0.016699699325297319 0 +402 0 -1.0857676 0.013649645 0.019827907448886111 0 +403 0 -1.1385398 0.01148938 0.016671629569056361 0 +405 0 -1.133943 0.01166328 0.016925452979744748 0 +407 0 -1.133943 0.01166328 0.016925452979744748 0 +408 0 -1.0594729 0.014870978 0.021615408626299879 0 +410 0 -1.133943 0.01166328 0.016925452979744748 0 411 0 ? ? ? 0 -412 1 1.2707446 0.9734032 0.038890554659704625 1 -417 0 -1.1213936 0.0079845805 0.011565549372462803 0 -420 0 -1.0793113 0.009247944 0.013404038471881364 0 -421 1 1.2902697 0.9751262 0.036339141410625192 1 -424 0 -1.1223897 0.007956849 0.011525220458022329 0 -425 1 1.3817995 0.9818543 0.026419111102529088 1 -426 0 -1.0906285 0.008889909 0.012882776063776632 0 -427 1 1.0748425 0.9483521 0.076505301165217018 1 -431 0 -1.1341074 0.0076377206 0.011061196710609239 0 -432 0 -1.1315845 0.0077053388 0.011159503328835802 0 -433 0 -1.0867381 0.009011409 0.013059647118945115 0 -435 1 1.1278855 0.95677006 0.063755846901879645 1 -437 0 -1.1246129 0.007895299 0.011435712878225906 0 -438 0 -1.09752 0.008678653 0.012575297643359769 0 -443 0 -1.1096194 0.0083197225 0.012053031127241418 0 -444 0 -1.0882932 0.008962646 0.012988658835760444 0 -445 0 -1.1203263 0.0080144 0.011608916132203563 0 -446 0 -1.1184047 0.008068365 0.011687403015130963 0 -447 0 -1.127014 0.007829355 0.011339821640851662 0 -448 0 -1.1069949 0.008396308 0.012164451831112489 0 -458 0 -1.1240406 0.007911098 0.011458687527139239 0 -459 0 -1.121116 0.0079923235 0.011576810139176335 0 -460 0 -1.1060008 0.008425498 0.012206921951198315 0 -461 0 -1.0908314 0.008883617 0.0128736172119411 0 -462 0 -1.1088237 0.008342869 0.012086704520405219 0 -463 0 -1.1261175 0.007853915 0.011375534142175974 0 -468 0 -1.1246129 0.007895299 0.011435712878225906 0 -469 0 -1.118102 0.008076902 0.011699818813457134 0 -470 0 -1.1144156 0.00818155 0.011852031423018961 0 -471 0 -1.1088237 0.008342869 0.012086704520405219 0 -472 0 -1.1190181 0.008051101 0.011662294062431083 0 -473 0 -1.1246129 0.007895299 0.011435712878225906 0 -475 0 -1.1223897 0.007956849 0.011525220458022329 0 -476 0 -1.1215473 0.007980294 0.011559314949213004 0 -477 0 -1.1246129 0.007895299 0.011435712878225906 0 -478 0 -1.1158155 0.008141652 0.011793997328444862 0 -479 1 1.0868475 0.95038354 0.073418239010687142 1 -481 0 -1.0707419 0.009528516 0.013812654636452324 0 -485 0 -1.069134 0.009582092 0.013890694439543166 0 -486 0 -1.1144156 0.00818155 0.011852031423018961 0 -488 1 0.82628876 0.88444036 0.17716322918289237 1 -490 0 -1.1184047 0.008068365 0.011687403015130963 0 -491 1 1.0258591 0.9392215 0.090462658823840608 1 -494 0 0.8485455 0.8922113 3.2137223945600253 1 -496 0 -1.1069949 0.008396308 0.012164451831112489 0 -498 0 -1.1276242 0.007812686 0.011315584066679825 0 -499 0 -1.1276242 0.007812686 0.011315584066679825 0 -500 0 -1.1357371 0.0075943535 0.010998151145919928 0 -503 0 -1.1306075 0.007731685 0.0111978084971893 0 -505 0 -1.0997922 0.008610096 0.012475527703355719 0 -506 1 1.2795259 0.974192 0.037721924504546818 1 -508 0 -1.127014 0.007829355 0.011339821640851662 0 -509 0 -1.1203263 0.0080144 0.011608916132203563 0 -511 0 -1.1300513 0.007746724 0.011219674408243712 0 -512 0 -1.127014 0.007829355 0.011339821640851662 0 -515 1 1.0804125 0.9493042 0.075057593767280467 1 -516 0 -1.1069949 0.008396308 0.012164451831112489 0 -518 0 -1.1085573 0.008350629 0.012097995099327711 0 -524 0 -1.1253841 0.00787406 0.011404828234943619 0 -525 0 -1.1116616 0.00826061 0.01196703631898092 0 -526 0 -1.1246129 0.007895299 0.011435712878225906 0 -530 1 1.0878738 0.95055366 0.073160030329738668 1 -536 0 -1.1330655 0.007665574 0.011101690705840397 0 -537 0 -1.1301197 0.0077448725 0.011216982451342549 0 -542 0 -1.1045994 0.00846682 0.012267044586532711 0 -543 0 -1.1276242 0.007812686 0.011315584066679825 0 -545 0 -1.1300513 0.007746724 0.011219674408243712 0 -550 0 -1.1253841 0.00787406 0.011404828234943619 0 -551 0 -1.124339 0.007902858 0.011446704476681715 0 -552 0 -1.1077493 0.00837422 0.012132315858355927 0 -553 0 -1.0724205 0.009472901 0.013731649751552989 0 -554 0 -1.1223897 0.007956849 0.011525220458022329 0 -555 0 -1.1217148 0.007975627 0.011552527950751233 0 -556 0 -1.0951291 0.008751375 0.012681134807253212 0 -562 0 -1.124339 0.007902858 0.011446704476681715 0 -564 0 -1.1320893 0.0076917633 0.011139766185253754 0 -567 0 -1.1131207 0.008218629 0.011905968376977249 0 -568 1 1.0061617 0.93513995 0.096745798377041115 1 -570 1 1.1159005 0.95499074 0.066441343726114849 1 -571 1 1.245299 0.9709828 0.042482369800438707 1 -572 0 -1.1253841 0.00787406 0.011404828234943619 0 -573 0 -1.1213936 0.0079845805 0.011565549372462803 0 -574 1 1.1540614 0.96042544 0.058254481592108789 1 -575 0 -1.1301197 0.0077448725 0.011216982451342549 0 -576 0 -1.1300513 0.007746724 0.011219674408243712 0 -579 0 -1.124339 0.007902858 0.011446704476681715 0 -580 0 -1.1270903 0.007827269 0.011336788197722327 0 -583 0 -1.1223897 0.007956849 0.011525220458022329 0 -585 0 -1.1184047 0.008068365 0.011687403015130963 0 -587 0 -1.1315845 0.0077053388 0.011159503328835802 0 -588 1 1.1140034 0.95470273 0.066876502658254414 1 -589 0 -1.127014 0.007829355 0.011339821640851662 0 -591 1 0.98675597 0.93087065 0.10334738080444947 1 -592 1 1.095488 0.95179856 0.071271825752004164 1 -595 0 -1.1300513 0.007746724 0.011219674408243712 0 -596 0 -1.124566 0.007896595 0.01143759672065953 0 -597 0 -1.1272281 0.007823502 0.011331310411443871 0 -598 0 -1.1253841 0.00787406 0.011404828234943619 0 -599 0 -1.0967449 0.008702163 0.012609511838485596 0 -601 0 -1.1100906 0.008306045 0.012033133397770807 0 -603 1 1.0804938 0.949318 0.075036669124115329 1 -605 1 1.2872839 0.9748699 0.036718385962829017 1 -608 1 1.2098006 0.9672463 0.048044798855875218 1 -610 1 1.2053971 0.9667515 0.048782973433250609 1 -611 1 1.1237099 0.95615786 0.064679265836639924 1 -615 0 -1.1215541 0.007980104 0.011559038646915988 0 -616 0 -1.1253841 0.00787406 0.011404828234943619 0 -620 0 -1.1253841 0.00787406 0.011404828234943619 0 -623 0 -1.1184047 0.008068365 0.011687403015130963 0 -625 0 -1.1042333 0.008477648 0.012282798922665665 0 -626 1 1.0391281 0.9418343 0.086454875290887995 1 -628 0 -1.1203263 0.0080144 0.011608916132203563 0 -630 0 -1.112858 0.008226169 0.011916936463632184 0 -631 0 -1.1300513 0.007746724 0.011219674408243712 0 -632 0 -1.1184047 0.008068365 0.011687403015130963 0 -635 0 -1.1130636 0.008220268 0.011908352736548146 0 -636 1 1.3525138 0.979923 0.029259690188593264 1 -637 0 -1.080772 0.009200939 0.013335593540205576 0 -640 0 -1.1033953 0.008502482 0.012318934539594937 0 -643 0 -1.1184047 0.008068365 0.011687403015130963 0 -646 0 -1.1038021 0.008490418 0.012301380208379595 0 -647 0 -1.0967114 0.008703181 0.012610993301778789 0 -648 1 1.373845 0.9813485 0.02716251101115931 1 -650 0 -1.0990745 0.008631693 0.012506957071952552 0 -651 0 -1.088386 0.008959743 0.012984432920237418 0 -655 0 -1.1253841 0.00787406 0.011404828234943619 0 -658 1 1.1447386 0.95915884 0.060158347918874477 1 -659 0 -1.1184047 0.008068365 0.011687403015130963 0 -662 0 -1.114841 0.008169406 0.011834367624079554 0 -663 0 -1.114841 0.008169406 0.011834367624079554 0 -664 0 -1.103993 0.00848476 0.012293147877896659 0 -666 0 -1.1101991 0.0083029 0.012028558015604242 0 -667 0 -1.1283454 0.007793028 0.01128700001539292 0 -669 1 1.1461244 0.9593495 0.059871576948251327 1 -671 0 -1.1063727 0.008414565 0.012191013971751003 0 -672 0 -1.1262618 0.007849955 0.011369775879992286 0 -673 0 -1.1120971 0.008248055 0.011948773649919505 0 -674 0 -1.1213936 0.0079845805 0.011565549372462803 0 -675 0 -1.11555 0.008149203 0.011804980819715493 0 -676 0 -1.118102 0.008076902 0.011699818813457134 0 -677 0 -1.127014 0.007829355 0.011339821640851662 0 -684 0 -1.1184047 0.008068365 0.011687403015130963 0 -686 0 -1.1184047 0.008068365 0.011687403015130963 0 -687 0 -1.1282966 0.007794355 0.01128892970540684 0 -690 0 -1.0967114 0.008703181 0.012610993301778789 0 -695 0 -1.1203263 0.0080144 0.011608916132203563 0 +412 1 1.3200305 0.9752503 0.036155552304814167 1 +417 0 -1.133943 0.01166328 0.016925452979744748 0 +420 0 -1.0521271 0.015231004 0.022142753454149094 0 +421 1 1.3525398 0.9777163 0.032512150576971682 1 +424 0 -1.138028 0.011508613 0.016699699325297319 0 +425 1 1.4302666 0.9826788 0.025208205479313395 1 +426 0 -1.0589776 0.014894988 0.021650571681855583 0 +427 1 1.1355786 0.9553847 0.065846364673318919 1 +431 0 -1.1028492 0.012909806 0.018746180407929646 0 +432 0 -1.1457884 0.011220354 0.016279048671876052 0 +433 0 -1.0514114 0.015266539 0.022194812906554104 0 +435 1 1.1902233 0.96248245 0.055167856610102504 1 +437 0 -1.1395046 0.011453209 0.016618840714141502 0 +438 0 -1.0617046 0.014763269 0.021457679839754298 0 +443 0 -1.1353163 0.011611057 0.016849224053613481 0 +444 0 -1.0006806 0.01800426 0.026211327565293416 0 +445 0 -1.136776 0.011555801 0.016768571834800274 0 +446 0 -1.1357223 0.011595662 0.016826752011794854 0 +447 0 -1.1412313 0.011388751 0.016524772114508274 0 +448 0 -1.1267003 0.011942566 0.017333189918116637 0 +458 0 -1.1320142 0.011737021 0.01703309713835971 0 +459 0 -1.1227953 0.012095876 0.017557059753949302 0 +460 0 -1.0821337 0.013812354 0.020065914373837184 0 +461 0 -1.1386843 0.011483956 0.016663713428360968 0 +462 0 -1.0842013 0.013719542 0.019930146152870601 0 +463 0 -1.1446584 0.011261879 0.016339637495810817 0 +468 0 -1.1395046 0.011453209 0.016618840714141502 0 +469 0 -1.1246219 0.012023925 0.017451988988607985 0 +470 0 -1.1279317 0.011894619 0.017263182221643397 0 +471 0 -1.0842013 0.013719542 0.019930146152870601 0 +472 0 -1.1347739 0.011631656 0.016879291543267477 0 +473 0 -1.1395046 0.011453209 0.016618840714141502 0 +475 0 -1.138028 0.011508613 0.016699699325297319 0 +476 0 -1.1302307 0.011805614 0.017133235794520311 0 +477 0 -1.1395046 0.011453209 0.016618840714141502 0 +478 0 -1.1230032 0.012087665 0.017545068104991856 0 +479 1 1.1334865 0.9550889 0.066293048640603924 1 +481 0 -1.0072789 0.01762258 0.025650695328507646 0 +485 0 -1.0681169 0.014458051 0.021010814794218919 0 +486 0 -1.1279317 0.011894619 0.017263182221643397 0 +488 1 0.83809525 0.88899153 0.16975841361651317 1 +490 0 -1.1357223 0.011595662 0.016826752011794854 0 +491 1 1.0891038 0.94835716 0.076497593855717333 1 +494 0 -0.8975685 0.025134355 0.036724692848664711 0 +496 0 -1.1267003 0.011942566 0.017333189918116637 0 +498 0 -1.1383487 0.011496558 0.016682105201821239 0 +499 0 -1.1383487 0.011496558 0.016682105201821239 0 +500 0 -1.1421459 0.011354756 0.016475163360988903 0 +503 0 -1.1373855 0.011532806 0.016735009192845141 0 +505 0 -1.1015399 0.01296509 0.018826982738233699 0 +506 1 1.294808 0.9731557 0.039257481114153066 1 +508 0 -1.1412313 0.011388751 0.016524772114508274 0 +509 0 -1.136776 0.011555801 0.016768571834800274 0 +511 0 -1.1403185 0.011422779 0.016574431502966412 0 +512 0 -1.1412313 0.011388751 0.016524772114508274 0 +515 1 1.1488951 0.9572242 0.063071238868351559 1 +516 0 -1.1267003 0.011942566 0.017333189918116637 0 +518 0 -1.0991306 0.013067421 0.018976562026547999 0 +524 0 -1.1366465 0.0115606915 0.016775709655013345 0 +525 0 -1.1084847 0.012674513 0.018402325629448959 0 +526 0 -1.1395046 0.011453209 0.016618840714141502 0 +530 1 1.1706043 0.9600686 0.058790624965559014 1 +536 0 -1.1396108 0.011449234 0.01661303838020094 0 +537 0 -1.1305658 0.011792697 0.017114377348216862 0 +542 0 -1.1047677 0.012829227 0.018628413943686733 0 +543 0 -1.1383487 0.011496558 0.016682105201821239 0 +545 0 -1.1403185 0.011422779 0.016574431502966412 0 +550 0 -1.1366465 0.0115606915 0.016775709655013345 0 +551 0 -1.1323191 0.011725332 0.017016033253838249 0 +552 0 -1.0833693 0.013756817 0.019984672064788322 0 +553 0 -1.0859339 0.013642245 0.019817083383789721 0 +554 0 -1.138028 0.011508613 0.016699699325297319 0 +555 0 -1.1419425 0.011362307 0.016486182548818219 0 +556 0 -1.0753529 0.014121086 0.020517630216033169 0 +562 0 -1.1323191 0.011725332 0.017016033253838249 0 +564 0 -1.1428149 0.011329953 0.016438969699364887 0 +567 0 -1.1399488 0.01143659 0.01659458632505377 0 +568 1 1.0746819 0.9459713 0.080131665390630644 1 +570 1 1.1933126 0.9628495 0.054617785659095808 1 +571 1 1.3151296 0.9748562 0.036738673945640045 1 +572 0 -1.1366465 0.0115606915 0.016775709655013345 0 +573 0 -1.133943 0.01166328 0.016925452979744748 0 +574 1 1.2059427 0.9643147 0.052424055438936738 1 +575 0 -1.1305658 0.011792697 0.017114377348216862 0 +576 0 -1.1403185 0.011422779 0.016574431502966412 0 +579 0 -1.1323191 0.011725332 0.017016033253838249 0 +580 0 -1.131185 0.011768859 0.017079576207995252 0 +583 0 -1.138028 0.011508613 0.016699699325297319 0 +585 0 -1.1357223 0.011595662 0.016826752011794854 0 +587 0 -1.1457884 0.011220354 0.016279048671876052 0 +588 1 1.1332173 0.95505077 0.066350672121087875 1 +589 0 -1.1412313 0.011388751 0.016524772114508274 0 +591 1 1.0596558 0.9433752 0.084096465459549199 1 +592 1 1.141254 0.9561777 0.064649318047402912 1 +595 0 -1.1403185 0.011422779 0.016574431502966412 0 +596 0 -1.1291512 0.011847323 0.017194128368617327 0 +597 0 -1.1215249 0.01214617 0.017630509248666527 0 +598 0 -1.1366465 0.0115606915 0.016775709655013345 0 +599 0 -1.0871316 0.013589063 0.019739298532038735 0 +601 0 -1.1249907 0.012009447 0.017430848462764999 0 +603 1 1.1388257 0.95584005 0.065158874886409604 1 +605 1 1.3509535 0.97760177 0.032681202690011096 1 +608 1 1.271825 0.9710976 0.0423118111615641 1 +610 1 1.2425501 0.9682535 0.046543293603521239 1 +611 1 1.1561115 0.95819056 0.061615494124720616 1 +615 0 -1.1199505 0.012208787 0.017721959456229827 0 +616 0 -1.1366465 0.0115606915 0.016775709655013345 0 +620 0 -1.1366465 0.0115606915 0.016775709655013345 0 +623 0 -1.1357223 0.011595662 0.016826752011794854 0 +625 0 -1.1292889 0.011841995 0.017186349411196268 0 +626 1 1.1126958 0.95204645 0.070896125080782663 1 +628 0 -1.136776 0.011555801 0.016768571834800274 0 +630 0 -1.1067946 0.012744632 0.018504787262103424 0 +631 0 -1.1403185 0.011422779 0.016574431502966412 0 +632 0 -1.1357223 0.011595662 0.016826752011794854 0 +635 0 -1.1100396 0.012610343 0.018308562512420464 0 +636 1 1.3834827 0.9798396 0.029382462061405611 1 +637 0 -1.0355865 0.016073387 0.023377380112382534 0 +640 0 -1.08037 0.013892011 0.02018244932566465 0 +643 0 -1.1357223 0.011595662 0.016826752011794854 0 +646 0 -1.1081604 0.01268794 0.018421945320872091 0 +647 0 -1.0766553 0.014061266 0.020430093679276402 0 +648 1 1.4006213 0.9809291 0.027779264694717205 1 +650 0 -1.1181357 0.012281361 0.01782795994044056 0 +651 0 -1.0568953 0.014996348 0.021799021296794223 0 +655 0 -1.1366465 0.0115606915 0.016775709655013345 0 +658 1 1.1925343 0.96275735 0.054755864301097634 1 +659 0 -1.1357223 0.011595662 0.016826752011794854 0 +662 0 -1.115283 0.0123963 0.017995853732251973 0 +663 0 -1.115283 0.0123963 0.017995853732251973 0 +664 0 -1.07764 0.014016198 0.020364149053015354 0 +666 0 -1.0908705 0.013424355 0.019498421483685031 0 +667 0 -1.1354455 0.011606155 0.016842068197582008 0 +669 1 1.1920625 0.9627014 0.054839736111358207 1 +671 0 -1.0879053 0.013554818 0.019689214085219193 0 +672 0 -1.1337652 0.011670057 0.016935344519875207 0 +673 0 -1.1038527 0.012867594 0.018684485920775835 0 +674 0 -1.133943 0.01166328 0.016925452979744748 0 +675 0 -1.1116662 0.012543555 0.01821098060593121 0 +676 0 -1.1246219 0.012023925 0.017451988988607985 0 +677 0 -1.1412313 0.011388751 0.016524772114508274 0 +684 0 -1.1357223 0.011595662 0.016826752011794854 0 +686 0 -1.1357223 0.011595662 0.016826752011794854 0 +687 0 -1.1469842 0.011176575 0.016215174089721913 0 +690 0 -1.0766553 0.014061266 0.020430093679276402 0 +695 0 -1.136776 0.011555801 0.016768571834800274 0 diff --git a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer-out.txt b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer-out.txt index 0d6047d595..bdeb0ffe66 100644 --- a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer-out.txt +++ b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer-out.txt @@ -11,43 +11,43 @@ Confusion table ||====================== PREDICTED || positive | negative | Recall TRUTH ||====================== - positive || 123 | 11 | 0.9179 - negative || 17 | 203 | 0.9227 + positive || 125 | 9 | 0.9328 + negative || 13 | 207 | 0.9409 ||====================== -Precision || 0.8786 | 0.9486 | -OVERALL 0/1 ACCURACY: 0.920904 -LOG LOSS/instance: 0.352056 +Precision || 0.9058 | 0.9583 | +OVERALL 0/1 ACCURACY: 0.937853 +LOG LOSS/instance: 0.309511 Test-set entropy (prior Log-Loss/instance): 0.956998 -LOG-LOSS REDUCTION (RIG): 0.632125 -AUC: 0.961330 +LOG-LOSS REDUCTION (RIG): 0.676582 +AUC: 0.971744 Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. TEST POSITIVE RATIO: 0.3191 (105.0/(105.0+224.0)) Confusion table ||====================== PREDICTED || positive | negative | Recall TRUTH ||====================== - positive || 104 | 1 | 0.9905 - negative || 44 | 180 | 0.8036 + positive || 102 | 3 | 0.9714 + negative || 22 | 202 | 0.9018 ||====================== -Precision || 0.7027 | 0.9945 | -OVERALL 0/1 ACCURACY: 0.863222 -LOG LOSS/instance: 0.139853 +Precision || 0.8226 | 0.9854 | +OVERALL 0/1 ACCURACY: 0.924012 +LOG LOSS/instance: 0.299203 Test-set entropy (prior Log-Loss/instance): 0.903454 -LOG-LOSS REDUCTION (RIG): 0.845202 -AUC: 0.993580 +LOG-LOSS REDUCTION (RIG): 0.668823 +AUC: 0.976488 OVERALL RESULTS --------------------------------------- -AUC: 0.977455 (0.0161) -Accuracy: 0.892063 (0.0288) -Positive precision: 0.790637 (0.0879) -Positive recall: 0.954193 (0.0363) -Negative precision: 0.971537 (0.0229) -Negative recall: 0.863149 (0.0596) -Log-loss: 0.245954 (0.1061) -Log-loss reduction: 0.738663 (0.1065) -F1 Score: 0.859972 (0.0378) -AUPRC: 0.964976 (0.0234) +AUC: 0.974116 (0.0024) +Accuracy: 0.930933 (0.0069) +Positive precision: 0.864189 (0.0416) +Positive recall: 0.952132 (0.0193) +Negative precision: 0.971850 (0.0135) +Negative recall: 0.921347 (0.0196) +Log-loss: 0.304357 (0.0052) +Log-loss reduction: 0.672702 (0.0039) +F1 Score: 0.904974 (0.0141) +AUPRC: 0.966608 (0.0115) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer-rp.txt b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer-rp.txt index 0a55cda83b..71dce23763 100644 --- a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer-rp.txt +++ b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer-rp.txt @@ -1,4 +1,4 @@ LdSvm AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC /bias /iter Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings -0.977455 0.892063 0.790637 0.954193 0.971537 0.863149 0.245954 0.738663 0.859972 0.964976 - 1000 LdSvm %Data% %Output% 99 0 0 maml.exe CV tr=LdSvm{iter=1000 bias=-} threads=- dout=%Output% data=%Data% seed=1 /bias:-;/iter:1000 +0.974116 0.930933 0.864189 0.952132 0.97185 0.921347 0.304357 0.672702 0.904974 0.966608 - 1000 LdSvm %Data% %Output% 99 0 0 maml.exe CV tr=LdSvm{iter=1000 bias=-} threads=- dout=%Output% data=%Data% seed=1 /bias:-;/iter:1000 diff --git a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer.txt b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer.txt index 64f3d10f02..393b27e5fe 100644 --- a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer.txt +++ b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-CV-breast-cancer.txt @@ -1,700 +1,700 @@ Instance Label Score Probability Log-loss Assigned -5 1 3.798577 0.9953265 0.0067582123624963198 1 -6 0 2.3537378 0.94675 4.2310742639508829 1 -8 0 -1.3237208 0.031004885 0.045438702609540707 0 -9 0 -1.1173705 0.043626778 0.064354358794450034 0 -10 0 -0.5647558 0.105482176 0.1608178653572119 0 -11 0 -0.9835599 0.05429453 0.080537153994703736 0 -18 1 2.4453433 0.95415115 0.067710263432171791 1 -20 1 1.6535794 0.8422055 0.24775575599557814 1 -21 1 1.5025896 0.80459064 0.31367313565393645 1 -25 1 2.2755082 0.9395503 0.089957724525288954 1 -28 0 -0.9835599 0.05429453 0.080537153994703736 0 -31 0 -1.1518178 0.041222338 0.060731797465077501 0 -32 1 -0.36576962 0.14236781 2.8123051114160096 0 -35 0 -0.9835599 0.05429453 0.080537153994703736 0 -37 0 -2.0270233 0.009463316 0.013717689090479346 0 +5 1 4.4443865 0.9968316 0.0045782976706429225 1 +6 0 1.9790748 0.8770795 3.0242024000497323 1 +8 0 -1.4800416 0.03397556 0.049868406130523811 0 +9 0 -1.1568757 0.054618265 0.081031103865112405 0 +10 0 -0.9528547 0.073245 0.10974009558551699 0 +11 0 -1.2762936 0.045885883 0.067766264933303177 0 +18 1 2.7628002 0.9596403 0.059434313892590365 1 +20 1 1.4911951 0.7713166 0.37460495707937808 1 +21 1 2.1454375 0.902081 0.14867109221656655 1 +25 1 2.450781 0.93640536 0.094794901326701 1 +28 0 -1.2762936 0.045885883 0.067766264933303177 0 +31 0 -1.5898459 0.028855124 0.042241561764939928 0 +32 1 0.028914407 0.26307422 1.9264582208267125 1 +35 0 -1.2762936 0.045885883 0.067766264933303177 0 +37 0 -2.5750813 0.006500593 0.0094089884132618398 0 40 0 ? ? ? 0 -41 1 -0.6496158 0.092491604 3.4345337739604118 0 -44 1 0.7855751 0.5456133 0.8740493120228453 1 -45 0 -0.8808264 0.06410698 0.095584464798131208 0 -46 1 4.658416 0.99892986 0.0015447149362973647 1 -48 0 -1.5890402 0.019877331 0.028965771777114586 0 -50 1 0.003942738 0.23860069 2.0673299027967555 1 -51 1 1.1339052 0.68602675 0.54366325869609766 1 -52 1 1.4008036 0.7756172 0.36658333105172886 1 -54 1 -0.33420587 0.14912057 2.7454488226576004 0 -56 1 4.244016 0.997821 0.0031471009300495095 1 -60 1 1.874571 0.8864047 0.17396257503832382 1 -63 1 0.30521545 0.34466052 1.5367520415071396 1 -64 0 -0.7521924 0.0787194 0.11828746108840639 0 -66 0 -1.6264237 0.018663496 0.027180167692602199 0 -68 1 1.6433578 0.8398569 0.25178452247065686 1 -69 0 -0.7168925 0.08323312 0.12537316697923573 0 -70 0 -1.4789128 0.023919972 0.034928656365260449 0 -71 1 -0.5153209 0.11377147 3.1357892908225713 0 -72 0 -0.31645623 0.1530327 0.23962183314951072 0 -73 1 2.7852516 0.9739069 0.038144267029598375 1 -74 1 0.07656033 0.26200673 1.9323242259342632 1 -76 0 -0.45335868 0.12495807 0.19257594257329302 0 -77 0 -0.9531292 0.057043348 0.084736642833019127 0 -79 0 -1.3471982 0.029815335 0.043668718542929044 0 -82 0 -1.513416 0.022573892 0.032940454043389661 0 -88 0 -1.6264237 0.018663496 0.027180167692602199 0 -90 0 -0.9194482 0.060237955 0.089632593834589547 0 -91 0 -1.0314138 0.050223183 0.074339552677276086 0 -92 0 -1.6264237 0.018663496 0.027180167692602199 0 -93 0 -0.7521924 0.0787194 0.11828746108840639 0 -95 0 -0.9194482 0.060237955 0.089632593834589547 0 -96 0 -0.79719615 0.073289804 0.1098098503118794 0 -97 0 -1.8683892 0.012392517 0.017990327475632902 0 -98 1 0.15964594 0.2905381 1.7832007177563522 1 -99 1 3.2506187 0.9881022 0.017267829986659239 1 -100 1 -0.6655141 0.09022358 3.4703516480128149 0 -102 0 -1.5208266 0.022294585 0.032528251233595692 0 -104 1 -0.5748169 0.10386175 3.267263667210329 0 -105 1 -0.97451854 0.055097938 4.1818578726685534 0 -106 1 4.267536 0.99790716 0.0030224913156800389 1 -108 0 -0.34790304 0.14615831 0.22795948527090723 0 -109 1 3.0392683 0.9829801 0.024765925116275084 1 -111 1 1.7172623 0.85621196 0.2239601063886936 1 -112 1 1.7306498 0.85902137 0.21923408026027552 1 -113 1 0.7023771 0.5099927 0.97145144540186357 1 -115 0 0.8394945 0.5684743 1.2124815873719856 1 -117 1 2.7680914 0.97314686 0.03927055895374601 1 -120 0 -0.7945155 0.073603325 0.11029801900101759 0 -121 0 -0.88858473 0.06331164 0.094358952168397833 0 -122 1 1.7416414 0.8612936 0.21542296049044318 1 -123 1 1.0231357 0.64364916 0.6356535739484932 1 -125 0 -0.7521924 0.0787194 0.11828746108840639 0 -128 1 2.4395082 0.95371044 0.068376789635910401 1 -129 0 -3.5990748 0.0006405246 0.00092437775262768705 0 -131 0 -1.1518178 0.041222338 0.060731797465077501 0 -132 1 4.2929554 0.9979964 0.0028934980943845895 1 -133 0 -1.0122521 0.05181752 0.076763361121481494 0 -137 0 -1.0486622 0.04882787 0.072221653693634799 0 -138 0 -1.3542684 0.029465852 0.043149120646902642 0 -141 0 -0.81492585 0.07124701 0.10663314686937736 0 -144 0 -0.9835599 0.05429453 0.080537153994703736 0 +41 1 -0.21146944 0.19793686 2.3368877776391721 0 +44 1 1.4779656 0.76771295 0.38136110946642249 1 +45 0 -1.2287059 0.04919396 0.072777026369907655 0 +46 1 5.1233683 0.9988809 0.0016153911773755033 1 +48 0 -2.1313438 0.012769605 0.018541281849617813 0 +50 1 0.15469992 0.30219522 1.7264472485018509 1 +51 1 1.2096398 0.6863995 0.54287955209755079 1 +52 1 1.898363 0.8630788 0.21243585499483142 1 +54 1 -0.49994987 0.13678004 2.8700703945831241 0 +56 1 4.7463503 0.99800503 0.0028810043736410378 1 +60 1 2.2085292 0.91031295 0.13556548949032204 1 +63 1 0.694081 0.4978826 1.006122484579141 1 +64 0 -1.0991795 0.05937848 0.088313753599299563 0 +66 0 -1.9247562 0.017454442 0.025403791535624963 0 +68 1 2.020052 0.883705 0.1783632153984685 1 +69 0 -1.1926565 0.051849224 0.076811598039861262 0 +70 0 -1.5320537 0.031449143 0.046100290311642825 0 +71 1 0.062300317 0.27313474 1.8723152822428462 1 +72 0 -1.291953 0.04484439 0.066192303116181517 0 +73 1 3.2780223 0.98129386 0.027242865992362034 1 +74 1 0.47862908 0.41596332 1.2654717723164717 1 +76 0 -0.82709277 0.08748552 0.13208064600488226 0 +77 0 -0.8628542 0.083199084 0.12531960905039682 0 +79 0 -1.7378896 0.02312244 0.033750344812203723 0 +82 0 -1.5865861 0.028995749 0.042450482613805018 0 +88 0 -1.9247562 0.017454442 0.025403791535624963 0 +90 0 -1.4114796 0.03760644 0.055301107076379863 0 +91 0 -1.2258475 0.049399707 0.073089249278292606 0 +92 0 -1.9247562 0.017454442 0.025403791535624963 0 +93 0 -1.0991795 0.05937848 0.088313753599299563 0 +95 0 -1.4114796 0.03760644 0.055301107076379863 0 +96 0 -1.0595361 0.06287173 0.093681566527410373 0 +97 0 -2.0817876 0.013765539 0.019997430586494856 0 +98 1 0.55291665 0.44391853 1.1716331765832924 1 +99 1 3.7987552 0.99150527 0.01230766120322987 1 +100 1 -0.14226519 0.21535334 2.2152223964064497 0 +102 0 -1.8089243 0.020782249 0.030298383897449148 0 +104 1 0.004260521 0.25579977 1.9669131258497021 1 +105 1 -0.8846363 0.08068267 3.6315973167124307 0 +106 1 4.605791 0.9975255 0.0035743540473884838 1 +108 0 -0.6445142 0.11261327 0.17236511272822935 0 +109 1 3.5597553 0.98778397 0.017732540655042602 1 +111 1 2.012019 0.8824311 0.18044447392317789 1 +112 1 1.795191 0.84325457 0.24595986948194126 1 +113 1 1.0815172 0.64257675 0.63805930418247436 1 +115 0 0.5099229 0.4276845 0.80511737950145801 1 +117 1 3.0094497 0.97201055 0.040956117980647883 1 +120 0 -1.4306396 0.0365558 0.053726983697238981 0 +121 0 -1.0751067 0.061477397 0.09153660356938273 0 +122 1 2.1498854 0.9026828 0.14770900681055196 1 +123 1 0.92897093 0.5871706 0.76814835935336534 1 +125 0 -1.0991795 0.05937848 0.088313753599299563 0 +128 1 2.7724514 0.9602105 0.058577379559975565 1 +129 0 -4.741499 0.00023478994 0.00033877005733192416 0 +131 0 -1.5898459 0.028855124 0.042241561764939928 0 +132 1 4.828172 0.9982402 0.0025411305149810205 1 +133 0 -1.2052909 0.050903548 0.075373385519932889 0 +137 0 -1.1569977 0.054608595 0.081016345783080257 0 +138 0 -1.5048554 0.03274673 0.048034391658362416 0 +141 0 -0.9856509 0.06989856 0.104540025896505 0 +144 0 -1.2762936 0.045885883 0.067766264933303177 0 145 0 ? ? ? 0 -147 0 -0.7625278 0.077440806 0.11628661294429539 0 -150 0 -0.5647558 0.105482176 0.1608178653572119 0 -151 1 0.80713165 0.55478156 0.85000826923535333 1 -152 1 3.0744305 0.9839621 0.023325320301425451 1 -154 0 -0.58416104 0.10237654 0.15581771742951545 0 -156 0 -0.3327535 0.14943755 0.23351092400175502 0 -161 0 -1.4098545 0.026853306 0.039270799193884563 0 +147 0 -1.2697616 0.0463271 0.068433572177866622 0 +150 0 -0.9528547 0.073245 0.10974009558551699 0 +151 1 0.8708193 0.5653662 0.82274243776960776 1 +152 1 3.8476553 0.992115 0.011420705879083178 1 +154 0 -0.8091333 0.089712694 0.13560613297330606 0 +156 0 -0.59574366 0.12031843 0.18494670428905347 0 +161 0 -1.6648756 0.025795508 0.037703459771881262 0 164 0 ? ? ? 0 -167 1 -1.0512761 0.048619654 4.362316562379803 0 -169 0 -0.21466784 0.17710659 0.28122252389339053 0 -171 0 -0.9194482 0.060237955 0.089632593834589547 0 -173 1 3.7376502 0.99481326 0.0075023523734084295 1 -174 1 2.3963907 0.9503268 0.0735043791656875 1 -176 0 -1.1518178 0.041222338 0.060731797465077501 0 -177 1 1.7295825 0.8587991 0.21960741572897097 1 -179 1 1.3639871 0.7644137 0.38757443283025766 1 -180 0 -0.5647558 0.105482176 0.1608178653572119 0 -181 0 -0.58416104 0.10237654 0.15581771742951545 0 -183 1 4.5968814 0.99881065 0.0017168921039059537 1 -187 1 1.370526 0.76643145 0.38377133098783217 1 -188 1 3.9094634 0.9961343 0.005587860923939124 1 -189 0 -0.37506866 0.14042762 0.21830896711257061 0 -191 1 2.8814652 0.97779536 0.032395531997584319 1 -192 0 -1.6957855 0.016600993 0.024151195819615917 0 -196 0 1.6869826 0.8496857 2.7339461037784374 1 -198 0 -0.58416104 0.10237654 0.15581771742951545 0 -199 0 -1.2180588 0.036950625 0.054318329476739667 0 -201 1 0.68633115 0.5030998 0.99108348120596257 1 -202 0 -0.9194482 0.060237955 0.089632593834589547 0 -204 0 -0.9194482 0.060237955 0.089632593834589547 0 -205 1 4.9666853 0.9993697 0.00090964497987452259 1 -206 1 1.8098085 0.8747065 0.19312906915894218 1 -207 0 -0.5647558 0.105482176 0.1608178653572119 0 -209 0 -1.4913768 0.023424912 0.034197119317690275 0 -210 1 4.5331125 0.9986731 0.0019156100973466857 1 -211 1 4.0517645 0.9969704 0.0043774012966327413 1 -212 0 -0.9194482 0.060237955 0.089632593834589547 0 -216 0 -0.7521924 0.0787194 0.11828746108840639 0 -218 1 2.801786 0.9746194 0.037089171833379952 1 -219 0 -1.881429 0.0121212145 0.017594063534513541 0 -223 1 2.1449943 0.9254828 0.11172190200006388 1 -226 1 2.922954 0.97929186 0.030189207611331086 1 -228 0 -0.5647558 0.105482176 0.1608178653572119 0 -233 1 1.2252245 0.71880704 0.47632355290562722 1 -237 1 1.2916322 0.7412895 0.43189102538203589 1 -239 1 1.5612386 0.8199535 0.28638599668698433 1 -240 0 -0.49985504 0.11647909 0.17866381811372251 0 -241 0 -0.9730798 0.05522681 0.081960069612246791 0 -242 0 -1.1518178 0.041222338 0.060731797465077501 0 -244 0 -0.9194482 0.060237955 0.089632593834589547 0 -246 1 4.7456794 0.99907875 0.0013296945100817375 1 -247 1 2.2552004 0.9375374 0.093051894539299823 1 -248 0 -0.46547523 0.12269885 0.1888559367183765 0 +167 1 -0.83620924 0.08637421 3.533255607336272 0 +169 0 -0.460068 0.14417431 0.22461110468072878 0 +171 0 -1.4114796 0.03760644 0.055301107076379863 0 +173 1 4.454449 0.99688005 0.0045081662175047825 1 +174 1 2.689062 0.95501846 0.066399473806099188 1 +176 0 -1.5898459 0.028855124 0.042241561764939928 0 +177 1 2.2698483 0.9177104 0.12388910086667157 1 +179 1 1.2246065 0.6913261 0.53256173946419683 1 +180 0 -0.9528547 0.073245 0.10974009558551699 0 +181 0 -0.8091333 0.089712694 0.13560613297330606 0 +183 1 5.183951 0.9989803 0.0014718901782633548 1 +187 1 1.8478985 0.8536594 0.2282675430195733 1 +188 1 4.6619816 0.99772966 0.0032791333711300451 1 +189 0 -0.7596451 0.09611623 0.14579082564205487 0 +191 1 3.3570597 0.98339665 0.024154655286084195 1 +192 0 -1.7699854 0.022034792 0.032144954321165313 0 +196 0 2.1109397 0.8973003 3.2834961694313538 1 +198 0 -0.8091333 0.089712694 0.13560613297330606 0 +199 0 -1.4470936 0.035676137 0.052410345077086848 0 +201 1 1.0158433 0.61909336 0.69177111240555278 1 +202 0 -1.4114796 0.03760644 0.055301107076379863 0 +204 0 -1.4114796 0.03760644 0.055301107076379863 0 +205 1 5.618711 0.99947673 0.00075511542751064559 1 +206 1 2.2790248 0.9187685 0.12222666070732108 1 +207 0 -0.9528547 0.073245 0.10974009558551699 0 +209 0 -1.6154233 0.027774468 0.040637073172576019 0 +210 1 5.2391615 0.9990631 0.0013523312601899733 1 +211 1 4.701762 0.9978639 0.0030850533036984613 1 +212 0 -1.4114796 0.03760644 0.055301107076379863 0 +216 0 -1.0991795 0.05937848 0.088313753599299563 0 +218 1 3.5366793 0.9873488 0.018368267719847502 1 +219 0 -1.9575747 0.016610729 0.024165479152208742 0 +223 1 2.4427197 0.93566406 0.09593745956707346 1 +226 1 3.5241416 0.987106 0.018723041773598108 1 +228 0 -0.9528547 0.073245 0.10974009558551699 0 +233 1 1.4129097 0.7494212 0.41615134482708016 1 +237 1 1.8847501 0.8605893 0.21660314938104278 1 +239 1 1.9253427 0.86790216 0.20439568088669627 1 +240 0 -0.63400716 0.11423597 0.17500567865283342 0 +241 0 -1.4770803 0.03412515 0.050091824976457047 0 +242 0 -1.5898459 0.028855124 0.042241561764939928 0 +244 0 -1.4114796 0.03760644 0.055301107076379863 0 +246 1 5.387262 0.9992536 0.0010772714717291792 1 +247 1 2.5475624 0.94470334 0.08206673522351246 1 +248 0 -0.4965086 0.13740526 0.21324517850777447 0 249 0 ? ? ? 0 -250 0 -0.17950663 0.18608594 0.29705162337149987 0 -252 0 2.7011049 0.96996886 5.0573967263792454 1 -254 1 1.6988889 0.85228056 0.23059967407267279 1 -257 0 -1.2180588 0.036950625 0.054318329476739667 0 -258 0 -1.387535 0.027874108 0.040784938030973716 0 -259 0 2.9103036 0.9788463 5.5629469511643306 1 -260 1 2.2986164 0.9417669 0.086558050207493395 1 -262 1 3.8893738 0.99599904 0.0057837457232200653 1 -267 1 2.2397575 0.935965 0.095473510577091622 1 -268 1 0.87979466 0.58537775 0.77256017757005047 1 -269 0 -0.9194482 0.060237955 0.089632593834589547 0 -271 0 -1.8683892 0.012392517 0.017990327475632902 0 -272 1 2.2397575 0.935965 0.095473510577091622 1 +250 0 -0.3374755 0.16899554 0.26707188241844787 0 +252 0 2.9809358 0.97079414 5.0975983864117129 1 +254 1 2.060449 0.88993096 0.16823467168007211 1 +257 0 -1.4470936 0.035676137 0.052410345077086848 0 +258 0 -1.7609001 0.022337493 0.032591567200563963 0 +259 0 3.213248 0.97937757 5.5996416940245259 1 +260 1 2.5005796 0.9408107 0.088023656675918119 1 +262 1 4.461219 0.9969122 0.0044616725823324608 1 +267 1 2.0278792 0.8849348 0.17635695803699203 1 +268 1 1.3706359 0.7370326 0.44019967678138688 1 +269 0 -1.4114796 0.03760644 0.055301107076379863 0 +271 0 -2.0817876 0.013765539 0.019997430586494856 0 +272 1 2.0278792 0.8849348 0.17635695803699203 1 275 0 ? ? ? 0 -276 0 -1.2180588 0.036950625 0.054318329476739667 0 -277 0 -0.7521924 0.0787194 0.11828746108840639 0 -278 0 -0.9194482 0.060237955 0.089632593834589547 0 -279 1 0.8271395 0.56325805 0.82813206270159867 1 -280 0 -1.387535 0.027874108 0.040784938030973716 0 -283 1 2.1588154 0.9271045 0.10919617313377075 1 -284 1 4.2253385 0.9977501 0.0032495715332427483 1 -285 1 3.3080983 0.9892093 0.01565229995818004 1 -288 1 0.9158053 0.6003151 0.73620815213206525 1 -290 0 -0.58416104 0.10237654 0.15581771742951545 0 -291 0 -0.9194482 0.060237955 0.089632593834589547 0 -293 1 2.6505725 0.96733314 0.047915272655849925 1 -296 0 1.4350817 0.7857031 2.2223172668755695 1 +276 0 -1.4470936 0.035676137 0.052410345077086848 0 +277 0 -1.0991795 0.05937848 0.088313753599299563 0 +278 0 -1.4114796 0.03760644 0.055301107076379863 0 +279 1 1.05752 0.63406837 0.65728968425813628 1 +280 0 -1.7609001 0.022337493 0.032591567200563963 0 +283 1 2.4556735 0.9368513 0.09410798306851309 1 +284 1 4.7114697 0.9978955 0.0030393810616290893 1 +285 1 3.967956 0.99343646 0.0095004051577309602 1 +288 1 1.1322452 0.66026497 0.59888299077412233 1 +290 0 -0.8091333 0.089712694 0.13560613297330606 0 +291 0 -1.4114796 0.03760644 0.055301107076379863 0 +293 1 2.994828 0.97139305 0.041872932337373034 1 +296 0 1.6395465 0.80901194 2.3884456179716924 1 297 0 ? ? ? 0 -299 1 1.3301523 0.75378144 0.40778182630733106 1 -300 1 1.7610326 0.8652272 0.20884903769525742 1 -301 0 -0.9194482 0.060237955 0.089632593834589547 0 -303 0 -0.9194482 0.060237955 0.089632593834589547 0 -304 1 2.5413437 0.96085155 0.057614540553021995 1 -308 1 1.6569057 0.84296376 0.24645749387094706 1 -309 0 -0.12644942 0.2002935 0.32245747398252661 0 -311 0 -0.58416104 0.10237654 0.15581771742951545 0 -312 1 -1.4542599 0.024929406 5.3260076750141074 0 -314 0 -0.40094522 0.13514486 0.20946958717752032 0 -316 1 2.5968194 0.9642845 0.052469267116382973 1 -317 1 3.517019 0.99243987 0.010948405787108205 1 -319 0 0.18599196 0.29995897 0.51448862126560058 1 +299 1 1.2973772 0.71465266 0.48468587543937386 1 +300 1 2.160904 0.90415925 0.14535120069128216 1 +301 0 -1.4114796 0.03760644 0.055301107076379863 0 +303 0 -1.4114796 0.03760644 0.055301107076379863 0 +304 1 2.7033174 0.9559496 0.064993530277659473 1 +308 1 2.203022 0.90962 0.13666414008025626 1 +309 0 -0.94577396 0.07398661 0.11089504364159686 0 +311 0 -0.8091333 0.089712694 0.13560613297330606 0 +312 1 -1.6879693 0.02491903 5.326608317838458 0 +314 0 -0.66283303 0.10983223 0.16785082382827582 0 +316 1 2.84118 0.96405315 0.052815401869740902 1 +317 1 4.1742563 0.9952103 0.006926692627745752 1 +319 0 -0.5427468 0.12920299 0.19959164389869499 0 321 0 ? ? ? 0 -323 1 3.3788874 0.9904335 0.013867961843596141 1 -327 0 -0.7521924 0.0787194 0.11828746108840639 0 -328 1 2.4722395 0.9561314 0.064719197189345101 1 -329 1 0.903454 0.5952112 0.74852640214198218 1 -331 0 -1.993043 0.010026717 0.014538503498786787 0 -332 0 -0.9593619 0.0564699 0.083859550198399982 0 -333 1 2.9478974 0.98014355 0.028935040084938541 1 -336 1 3.3775115 0.9904111 0.013900607249589017 1 -338 0 -0.40094522 0.13514486 0.20946958717752032 0 -343 0 -0.58416104 0.10237654 0.15581771742951545 0 -344 1 0.9923593 0.6314274 0.66331121233792234 1 -346 0 -0.9282099 0.059391152 0.088333191841281397 0 -347 0 0.14850712 0.28660804 0.48723313898397874 1 -348 1 0.0134409685 0.24157889 2.0494336912177311 1 -349 1 2.1005151 0.9200365 0.12023700575631781 1 -350 0 -1.4897162 0.023490287 0.034293701258874813 0 -352 0 0.41824618 0.38975632 0.71254265061003341 1 -353 1 4.800415 0.9991614 0.0012103195019509362 1 -354 0 -0.7521924 0.0787194 0.11828746108840639 0 -355 0 -1.6184881 0.018914932 0.027549860272256466 0 -358 1 1.8149067 0.8756636 0.19155138811686021 1 -360 1 3.074907 0.98397505 0.023306356161828405 1 -361 1 3.5106673 0.9923575 0.011068156059149552 1 -366 1 4.079712 0.9971121 0.0041723933495482618 1 -368 0 -0.4016338 0.13500659 0.20923895614599311 0 -370 0 -1.0076947 0.0522037 0.07735106580622049 0 -371 0 -0.4016338 0.13500659 0.20923895614599311 0 -373 0 -1.679495 0.0170643 0.024831050151857808 0 -376 0 -0.7521924 0.0787194 0.11828746108840639 0 -377 0 -0.40094522 0.13514486 0.20946958717752032 0 -378 0 -1.3066629 0.031897858 0.046768823944684318 0 -379 0 -1.4920803 0.023397267 0.034156279358109824 0 -381 1 2.55193 0.96153027 0.056595821738627622 1 -383 0 -0.81492585 0.07124701 0.10663314686937736 0 -384 0 -0.81492585 0.07124701 0.10663314686937736 0 -387 0 -1.2122881 0.037305173 0.054849556086386461 0 -388 0 -0.78274155 0.074995026 0.11246697146726271 0 -389 0 -1.4642861 0.024513962 0.035806870630939736 0 -391 1 4.085717 0.99714166 0.0041296187554742819 1 -392 0 -1.2180588 0.036950625 0.054318329476739667 0 -395 0 -1.2180588 0.036950625 0.054318329476739667 0 -396 0 -1.387535 0.027874108 0.040784938030973716 0 -398 0 -0.6688784 0.08975011 0.13566543530967909 0 -399 0 -0.07644083 0.214415 0.34816070982399361 0 -404 0 0.0023503893 0.23810388 0.39233378943933139 1 -406 0 -1.0053054 0.052407246 0.077660928314294095 0 -409 0 -1.1198002 0.04345288 0.064092058595875398 0 -413 0 -1.7598194 0.014896766 0.021653175426926657 0 -414 1 3.0163555 0.98230857 0.025751814202236625 1 -415 0 1.0804753 0.6659187 1.5817288978314228 1 -416 1 1.1454417 0.6902815 0.53474325362775377 1 -418 0 -0.81253624 0.07151924 0.10705608112237171 0 -419 0 -1.2437469 0.035411324 0.052014222049020994 0 -422 0 -0.20523377 0.17948194 0.28539300371190612 0 -423 0 -1.4789128 0.023919972 0.034928656365260449 0 -428 0 -0.7521924 0.0787194 0.11828746108840639 0 -429 0 -0.9835599 0.05429453 0.080537153994703736 0 -430 0 0.10687516 0.27220482 0.45839559102139221 1 -434 0 3.031101 0.9827436 5.8567264113997481 1 -436 1 1.4567108 0.79189545 0.33661812535623753 1 -439 0 -1.5239791 0.022176793 0.032354448950258352 0 -440 1 0.9338661 0.6077389 0.71847642764733566 1 -441 0 0.2622591 0.328182 0.57385767014278766 1 -442 0 0.19004305 0.30142298 0.51750891650350606 1 -449 1 4.401811 0.9983377 0.0024002075773767819 1 -450 0 -1.0427724 0.04930014 0.072938149374641353 0 -451 0 -1.5239791 0.022176793 0.032354448950258352 0 -452 0 -1.0115172 0.051879615 0.076857841243832117 0 -453 1 2.9439082 0.9800097 0.029132102758451257 1 -454 0 -0.5226059 0.11251519 0.1722056660046673 0 -455 1 -0.261382 0.16570765 2.5932879069981074 0 -456 1 2.4491978 0.95444006 0.06727350153661063 1 -457 1 1.5999807 0.8295746 0.26955639879850685 1 -464 0 -1.2850319 0.03306604 0.048510738157621122 0 -465 1 1.9063592 0.8917906 0.16522305601593093 1 -466 1 2.490209 0.95740867 0.062793229272849041 1 -467 1 3.3257709 0.9895287 0.015186520744684141 1 -474 0 -1.5239791 0.022176793 0.032354448950258352 0 -480 0 -1.1561064 0.04093202 0.060295013354904448 0 -482 1 0.77131796 0.5395322 0.89021907182084137 1 -483 1 3.7463102 0.9948895 0.0073918002774604587 1 -484 0 -1.4241669 0.026217945 0.038329180114913208 0 -487 1 2.3907423 0.9498666 0.074203191232924973 1 -489 1 -0.5039825 0.11575107 3.1109025306484743 0 -492 0 -1.0886832 0.04573076 0.067531723169546795 0 -493 1 4.591501 0.9987996 0.0017328195302273959 1 -495 0 -0.8565059 0.06666089 0.099526741997412072 0 -497 0 -0.95630044 0.05675089 0.084289260899524363 0 -501 0 -1.4554874 0.024878178 0.036345628520759621 0 -502 0 -1.4382305 0.025607888 0.037425640140168727 0 -504 0 -0.58416104 0.10237654 0.15581771742951545 0 -507 0 0.47155622 0.41175452 0.7655097617997153 1 -510 0 -0.58416104 0.10237654 0.15581771742951545 0 -513 0 -0.8565059 0.06666089 0.099526741997412072 0 -514 1 4.615252 0.99884754 0.001663601074808342 1 -517 0 -0.40094522 0.13514486 0.20946958717752032 0 -519 1 2.957257 0.9804542 0.028477846057154331 1 -520 0 -0.93872654 0.058389466 0.08679763507242598 0 -521 0 -1.6552678 0.017776912 0.025877360280510533 0 -522 1 0.06758494 0.25903508 1.9487806020592569 1 -523 1 2.997628 0.9817405 0.026586399620747828 1 -527 0 -1.6264237 0.018663496 0.027180167692602199 0 -528 0 -1.5221341 0.022245657 0.032456055291985163 0 -529 0 -1.0886832 0.04573076 0.067531723169546795 0 -531 0 -1.0053054 0.052407246 0.077660928314294095 0 -532 0 -0.5647558 0.105482176 0.1608178653572119 0 -533 0 -1.2180588 0.036950625 0.054318329476739667 0 -534 0 -0.9835599 0.05429453 0.080537153994703736 0 -535 0 -0.6844344 0.08758982 0.13224555590438908 0 -538 0 -1.4554874 0.024878178 0.036345628520759621 0 -539 0 -1.9389408 0.010992989 0.015947346623104423 0 -540 0 -1.3213655 0.03112673 0.045620123681692756 0 -541 0 -1.0486622 0.04882787 0.072221653693634799 0 -544 0 -0.7538145 0.07851746 0.11797126352044145 0 -546 1 5.860985 0.9998644 0.00019564353034305721 1 -547 0 -0.46166164 0.123406105 0.19001946315440138 0 -548 0 -0.6906426 0.08674088 0.13090384107125297 0 -549 1 2.3525813 0.9466497 0.079097472779850853 1 -557 0 -1.4897162 0.023490287 0.034293701258874813 0 -558 0 -0.9835599 0.05429453 0.080537153994703736 0 -559 0 -1.6957855 0.016600993 0.024151195819615917 0 -560 0 -1.8683892 0.012392517 0.017990327475632902 0 -561 0 -1.8683892 0.012392517 0.017990327475632902 0 -563 0 -1.2180588 0.036950625 0.054318329476739667 0 -565 1 4.6540093 0.99892175 0.001556422332581576 1 -566 0 -1.2847495 0.033081565 0.048533899559077096 0 -569 1 1.2685928 0.73362356 0.44688811534508827 1 -577 0 -0.7521924 0.0787194 0.11828746108840639 0 -578 0 -0.7521924 0.0787194 0.11828746108840639 0 -581 1 3.8309803 0.9955785 0.006392980151636583 1 -582 1 5.2395725 0.9996056 0.00056911683871863062 1 -584 0 -1.8211826 0.013425774 0.019500497016280361 0 -586 1 4.6255355 0.99886775 0.0016344166766065324 1 -590 1 0.90911347 0.5975525 0.74286267572209164 1 -593 0 -1.4241669 0.026217945 0.038329180114913208 0 -594 1 3.278635 0.98865515 0.016460708619123424 1 -600 0 -1.2180588 0.036950625 0.054318329476739667 0 -602 0 -1.4554874 0.024878178 0.036345628520759621 0 -604 1 1.1635954 0.6969118 0.52095198784248442 1 -606 0 -1.2683991 0.033992298 0.049893403063598005 0 -607 0 -0.58416104 0.10237654 0.15581771742951545 0 -609 0 -1.5239791 0.022176793 0.032354448950258352 0 -612 1 3.5652463 0.993037 0.010080643341518368 1 -613 0 -0.32578808 0.15096553 0.2361049622524784 0 -614 0 -0.6271124 0.09578942 0.14526929263485608 0 +323 1 3.3992307 0.9844217 0.022651678164352436 1 +327 0 -1.0991795 0.05937848 0.088313753599299563 0 +328 1 3.2761424 0.98124075 0.027320946929866023 1 +329 1 1.0183318 0.6199942 0.68967332183637942 1 +331 0 -2.1959515 0.011577326 0.016799988838971942 0 +332 0 -0.9301793 0.075644374 0.11348009031009064 0 +333 1 3.094405 0.9753499 0.036008221750438135 1 +336 1 3.6950512 0.9900532 0.014422078825868609 1 +338 0 -0.66283303 0.10983223 0.16785082382827582 0 +343 0 -0.8091333 0.089712694 0.13560613297330606 0 +344 1 0.9751647 0.6042551 0.72677039750992734 1 +346 0 -0.80091536 0.090748765 0.1372491139744203 0 +347 0 0.03877846 0.26602173 0.44619074038254897 1 +348 1 -1.0478476 0.063937746 3.9671883044101328 0 +349 1 1.8514928 0.85434765 0.22710485239053277 1 +350 0 -1.6117336 0.027927902 0.040864772668825225 0 +352 0 0.32662937 0.36058822 0.64518277767678511 1 +353 1 5.233806 0.9990554 0.0013634345868463072 1 +354 0 -1.0991795 0.05937848 0.088313753599299563 0 +355 0 -1.8193836 0.020457853 0.029820525323652834 0 +358 1 2.5689664 0.9463957 0.079484583460322847 1 +360 1 3.7426846 0.99074835 0.01340944140796198 1 +361 1 3.9218044 0.9929577 0.010195818334764142 1 +366 1 4.7482243 0.99801075 0.0028727327284018694 1 +368 0 -0.65729076 0.11066721 0.16920470967957332 0 +370 0 -0.9821722 0.070246704 0.10508013752912485 0 +371 0 -0.65729076 0.11066721 0.16920470967957332 0 +373 0 -1.8460634 0.019652665 0.028635110770960329 0 +376 0 -1.0991795 0.05937848 0.088313753599299563 0 +377 0 -0.66283303 0.10983223 0.16785082382827582 0 +378 0 -1.976651 0.01613887 0.0234733988220137 0 +379 0 -1.5106086 0.03246801 0.047618731897701709 0 +381 1 3.241781 0.9802443 0.02878677825818967 1 +383 0 -0.9856509 0.06989856 0.104540025896505 0 +384 0 -0.9856509 0.06989856 0.104540025896505 0 +387 0 -1.2669839 0.04651594 0.068719277050113894 0 +388 0 -1.0470998 0.06400652 0.095429618197872074 0 +389 0 -1.5950111 0.028633654 0.041912591508654735 0 +391 1 4.779634 0.9981042 0.0027376359023654488 1 +392 0 -1.4470936 0.035676137 0.052410345077086848 0 +395 0 -1.4470936 0.035676137 0.052410345077086848 0 +396 0 -1.7609001 0.022337493 0.032591567200563963 0 +398 0 -0.7441591 0.09820246 0.14912451963547385 0 +399 0 -0.2396788 0.19114862 0.30605345842765247 0 +404 0 -0.28313833 0.18104103 0.28813692416311421 0 +406 0 -1.0388905 0.064766034 0.096600768747693563 0 +409 0 -1.3499125 0.041182328 0.060671595319014449 0 +413 0 -1.9563836 0.016640637 0.024209357494375881 0 +414 1 3.4534392 0.9856483 0.020855177771726248 1 +415 0 0.815082 0.5442259 1.1336090611022984 1 +416 1 1.3912925 0.74313504 0.42830370824363628 1 +418 0 -0.90365964 0.078541964 0.1180096295796875 0 +419 0 -1.4326301 0.03644829 0.053566000632826398 0 +422 0 -0.12315461 0.22035432 0.35910947072130694 0 +423 0 -1.5320537 0.031449143 0.046100290311642825 0 +428 0 -1.0991795 0.05937848 0.088313753599299563 0 +429 0 -1.2762936 0.045885883 0.067766264933303177 0 +430 0 -0.0003312072 0.25445962 0.42364160000772078 0 +434 0 3.3855588 0.98409635 5.9744981220474855 1 +436 1 1.945739 0.87145233 0.19850634382087728 1 +439 0 -1.4834489 0.03380423 0.049612558800140524 0 +440 1 0.21734804 0.32286215 1.6310097827140879 1 +441 0 0.15582596 0.30256003 0.51985905106456065 1 +442 0 -0.32195094 0.17237045 0.27294293598655706 0 +449 1 5.0136595 0.9986758 0.0019116492460795706 1 +450 0 -1.3075589 0.043828882 0.064659266849870944 0 +451 0 -1.4834489 0.03380423 0.049612558800140524 0 +452 0 -1.3045118 0.044025425 0.064955845738404697 0 +453 1 3.100035 0.9755569 0.035702058614062906 1 +454 0 -0.5667803 0.12510663 0.19282090406330452 0 +455 1 0.049559847 0.2692673 1.8928891053838652 1 +456 1 3.0048475 0.9718176 0.041242515643025127 1 +457 1 2.223891 0.9122206 0.13254534933574869 1 +464 0 -1.3228716 0.042853814 0.063188808356615042 0 +465 1 2.6221776 0.9503935 0.073403128820453856 1 +466 1 3.0790818 0.9747777 0.036854850193771896 1 +467 1 3.781155 0.99127454 0.012643424581591636 1 +474 0 -1.4834489 0.03380423 0.049612558800140524 0 +480 0 -1.0657641 0.062310517 0.092817843003226097 0 +482 1 1.464867 0.76410615 0.38815501441674094 1 +483 1 4.4291334 0.99675673 0.0046866501356489544 1 +484 0 -1.3796455 0.0394165 0.058017066510644429 0 +487 1 2.792164 0.9613512 0.056864499703949675 1 +489 1 -0.5502564 0.12791091 2.966788749588904 0 +492 0 -1.1693292 0.053639047 0.079537545571283952 0 +493 1 5.070708 0.9987868 0.0017513300033648797 1 +495 0 -1.0054252 0.067949735 0.10152033351944299 0 +497 0 -0.8955821 0.07944451 0.11942341100470992 0 +501 0 -1.6115539 0.027935391 0.040875888499771439 0 +502 0 -1.692199 0.024761675 0.036173272550836286 0 +504 0 -0.8091333 0.089712694 0.13560613297330606 0 +507 0 0.14653797 0.29955843 0.51366338898533981 1 +510 0 -0.8091333 0.089712694 0.13560613297330606 0 +513 0 -1.0054252 0.067949735 0.10152033351944299 0 +514 1 5.258286 0.9990902 0.0013131690460845696 1 +517 0 -0.66283303 0.10983223 0.16785082382827582 0 +519 1 3.1457021 0.97717476 0.033311496595379603 1 +520 0 -1.2451435 0.048026484 0.071006655969873739 0 +521 0 -1.7643654 0.022221562 0.0324205028393428 0 +522 1 0.24495678 0.33220097 1.5898717906751076 1 +523 1 3.237643 0.98012084 0.028968466897409107 1 +527 0 -1.9247562 0.017454442 0.025403791535624963 0 +528 0 -1.4455717 0.03575664 0.05253078893051133 0 +529 0 -1.1693292 0.053639047 0.079537545571283952 0 +531 0 -1.0388905 0.064766034 0.096600768747693563 0 +532 0 -0.9528547 0.073245 0.10974009558551699 0 +533 0 -1.4470936 0.035676137 0.052410345077086848 0 +534 0 -1.2762936 0.045885883 0.067766264933303177 0 +535 0 -1.3204519 0.043006506 0.063418977848035019 0 +538 0 -1.6115539 0.027935391 0.040875888499771439 0 +539 0 -1.9228759 0.017504035 0.025476611849228011 0 +540 0 -1.4221979 0.037015174 0.05441502893407428 0 +541 0 -1.1569977 0.054608595 0.081016345783080257 0 +544 0 -1.1552584 0.054746665 0.081227060706948087 0 +546 1 6.5617037 0.999877 0.00017749701410946082 1 +547 0 -0.56898683 0.12473619 0.19221017480718539 0 +548 0 -0.7386718 0.098951325 0.15032305150863456 0 +549 1 3.020277 0.9724594 0.040290111597698054 1 +557 0 -1.6117336 0.027927902 0.040864772668825225 0 +558 0 -1.2762936 0.045885883 0.067766264933303177 0 +559 0 -1.7699854 0.022034792 0.032144954321165313 0 +560 0 -2.0817876 0.013765539 0.019997430586494856 0 +561 0 -2.0817876 0.013765539 0.019997430586494856 0 +563 0 -1.4470936 0.035676137 0.052410345077086848 0 +565 1 5.2710276 0.9991078 0.001287778728059478 1 +566 0 -1.6558733 0.026145246 0.038221477750395964 0 +569 1 1.7556311 0.8350554 0.26005616278158594 1 +577 0 -1.0991795 0.05937848 0.088313753599299563 0 +578 0 -1.0991795 0.05937848 0.088313753599299563 0 +581 1 4.179206 0.99524635 0.0068744184401775543 1 +582 1 6.0436506 0.9997275 0.00039320591905750955 1 +584 0 -2.4150288 0.008297022 0.012020007495764925 0 +586 1 5.416837 0.9992867 0.0010294253738857779 1 +590 1 1.0599657 0.63493943 0.65530911735345609 1 +593 0 -1.3796455 0.0394165 0.058017066510644429 0 +594 1 3.5707138 0.9879854 0.017438412505960923 1 +600 0 -1.4470936 0.035676137 0.052410345077086848 0 +602 0 -1.6115539 0.027935391 0.040875888499771439 0 +604 1 -0.3457305 0.16722253 2.5801588621559017 0 +606 0 -1.3867486 0.039005518 0.057399947713308599 0 +607 0 -0.8091333 0.089712694 0.13560613297330606 0 +609 0 -1.4834489 0.03380423 0.049612558800140524 0 +612 1 4.2750416 0.9958943 0.0059354473781710396 1 +613 0 -0.4090018 0.15412405 0.24148199343902133 0 +614 0 -0.83644336 0.08634584 0.13027991543500039 0 617 0 ? ? ? 0 -618 0 -1.4554874 0.024878178 0.036345628520759621 0 -619 0 -1.6957855 0.016600993 0.024151195819615917 0 -621 0 -1.3970283 0.027435396 0.040134008318457183 0 -622 0 -2.0290043 0.009431454 0.013671284292189386 0 -624 0 -1.3024747 0.03212088 0.047101216282497582 0 -627 0 0.6356638 0.48133904 0.94713630665753967 1 -629 0 -1.2850319 0.03306604 0.048510738157621122 0 -633 1 1.6099855 0.8319918 0.26535879980924404 1 -634 0 -1.0486622 0.04882787 0.072221653693634799 0 -638 0 -1.2850319 0.03306604 0.048510738157621122 0 -639 0 -1.4897162 0.023490287 0.034293701258874813 0 -641 0 -1.2180588 0.036950625 0.054318329476739667 0 -642 0 -1.2180588 0.036950625 0.054318329476739667 0 -644 0 -0.81492585 0.07124701 0.10663314686937736 0 -645 0 -1.2180588 0.036950625 0.054318329476739667 0 -649 0 -1.2180588 0.036950625 0.054318329476739667 0 -652 0 -1.480278 0.023865255 0.03484778407822689 0 -653 0 -1.4554874 0.024878178 0.036345628520759621 0 -654 0 -1.387535 0.027874108 0.040784938030973716 0 -656 0 -1.6957855 0.016600993 0.024151195819615917 0 -657 0 -1.092976 0.045409884 0.067046696806395148 0 -660 0 -0.7521924 0.0787194 0.11828746108840639 0 -661 0 -1.6264237 0.018663496 0.027180167692602199 0 -665 0 -0.58416104 0.10237654 0.15581771742951545 0 -668 1 0.5584668 0.44834724 1.157311576062372 1 -670 1 4.044202 0.9969309 0.0044345879270996803 1 -678 0 -0.58416104 0.10237654 0.15581771742951545 0 -679 0 -0.81492585 0.07124701 0.10663314686937736 0 -680 1 3.831054 0.9955791 0.0063921164196725127 1 -681 1 5.3066554 0.9996485 0.00050717998528051805 1 -682 0 -1.6471527 0.018022083 0.026237513347803065 0 -683 0 -0.58416104 0.10237654 0.15581771742951545 0 -685 0 -0.58416104 0.10237654 0.15581771742951545 0 -688 0 -1.2850319 0.03306604 0.048510738157621122 0 -689 0 -1.469554 0.024298396 0.0354880948406118 0 -691 1 3.4326797 0.9912709 0.012648716234188933 1 -692 0 -1.0486622 0.04882787 0.072221653693634799 0 -693 0 -1.3447275 0.02993841 0.043851745465984887 0 -694 0 -1.0603099 0.047906514 0.070824856652635659 0 -696 1 2.3973944 0.95040816 0.073380870984176727 1 -697 1 2.3173213 0.94350505 0.083897857093712419 1 -698 1 2.2249517 0.9344229 0.097852448188518584 1 -0 0 -0.66704065 0.0025475002 0.0036799552259266476 0 -1 0 1.8861008 0.8152147 2.4360780521440093 1 -2 0 -0.31498727 0.007088205 0.010262532567395727 0 -3 0 2.217634 0.92072415 3.6569748127418911 1 -4 0 -0.44860676 0.0048095263 0.0069554192612883122 0 -7 0 -0.7394654 0.002062959 0.0029792947106998661 0 -12 1 1.36338 0.4895217 1.0305552448776814 1 -13 0 -0.10010879 0.013192403 0.019159272718974336 0 -14 1 2.814325 0.98514384 0.02159370782527905 1 -15 1 1.2049481 0.37648627 1.409330836171141 1 -16 0 -0.57446826 0.0033354373 0.0048200617941456031 0 -17 0 -0.7354426 0.002087281 0.0030144571763673074 0 -19 0 -0.5865108 0.0032205714 0.0046538004195265494 0 -22 0 -0.6272879 0.002860122 0.004132195786456465 0 +618 0 -1.6115539 0.027935391 0.040875888499771439 0 +619 0 -1.7699854 0.022034792 0.032144954321165313 0 +621 0 -1.9508606 0.016780011 0.024413849208705655 0 +622 0 -2.062994 0.014162944 0.020578883874484137 0 +624 0 -1.3873832 0.038969 0.057345124496166265 0 +627 0 0.37153038 0.3766359 0.68185304442277472 1 +629 0 -1.3228716 0.042853814 0.063188808356615042 0 +633 1 2.2955484 0.9206425 0.11928705865304566 1 +634 0 -1.1569977 0.054608595 0.081016345783080257 0 +638 0 -1.3228716 0.042853814 0.063188808356615042 0 +639 0 -1.6117336 0.027927902 0.040864772668825225 0 +641 0 -1.4470936 0.035676137 0.052410345077086848 0 +642 0 -1.4470936 0.035676137 0.052410345077086848 0 +644 0 -0.9856509 0.06989856 0.104540025896505 0 +645 0 -1.4470936 0.035676137 0.052410345077086848 0 +649 0 -1.4470936 0.035676137 0.052410345077086848 0 +652 0 -1.5071157 0.03263695 0.047870664595627034 0 +653 0 -1.6115539 0.027935391 0.040875888499771439 0 +654 0 -1.7609001 0.022337493 0.032591567200563963 0 +656 0 -1.7699854 0.022034792 0.032144954321165313 0 +657 0 -0.9086896 0.07798469 0.11713738922560447 0 +660 0 -1.0991795 0.05937848 0.088313753599299563 0 +661 0 -1.9247562 0.017454442 0.025403791535624963 0 +665 0 -0.8091333 0.089712694 0.13560613297330606 0 +668 1 1.1277544 0.65871614 0.60227119034673204 1 +670 1 4.515658 0.9971592 0.0041042650584377652 1 +678 0 -0.8091333 0.089712694 0.13560613297330606 0 +679 0 -0.9856509 0.06989856 0.104540025896505 0 +680 1 4.591314 0.99747 0.0036546127974618854 1 +681 1 5.789462 0.9995974 0.00058098837263221594 1 +682 0 -1.8108897 0.02072091 0.030208015474333081 0 +683 0 -0.8091333 0.089712694 0.13560613297330606 0 +685 0 -0.8091333 0.089712694 0.13560613297330606 0 +688 0 -1.3228716 0.042853814 0.063188808356615042 0 +689 0 -1.6746339 0.025421549 0.037149771035776144 0 +691 1 3.7655423 0.99106467 0.012948897486997858 1 +692 0 -1.1569977 0.054608595 0.081016345783080257 0 +693 0 -1.5692985 0.029752707 0.043575592166996108 0 +694 0 -1.2795825 0.04566525 0.067432687600591998 0 +696 1 2.995239 0.9714106 0.041846906598692654 1 +697 1 2.7384136 0.95816445 0.061654802290764095 1 +698 1 2.5055556 0.9412348 0.08737339154713844 1 +0 0 -3.340107 2.394717E-05 3.4548877251977815E-05 0 +1 0 1.7701719 0.8694952 2.9378253684855622 1 +2 0 -0.8362533 0.011015482 0.015980158671046421 0 +3 0 2.0550754 0.93056536 3.8482005483853006 1 +4 0 -3.427366 1.9332694E-05 2.789145156796553E-05 0 +7 0 -2.6956668 0.00011635529 0.00016787496359475996 0 +12 1 0.9912622 0.4964356 1.010321482142972 1 +13 0 -0.4236125 0.02973863 0.043554659432650633 0 +14 1 2.6703978 0.9837754 0.023599148313491945 1 +15 1 0.18422095 0.119831294 3.0609233792044646 1 +16 0 -1.1420736 0.0052326466 0.0075689333865708005 0 +17 0 -3.0617628 4.7400794E-05 6.8386511016753427E-05 0 +19 0 -3.6175222 1.2125714E-05 1.7493814123837223E-05 0 +22 0 -1.0734073 0.006186717 0.008953270531768193 0 23 1 ? ? ? 0 -24 0 -0.8497671 0.0014958045 0.0021596053416366974 0 -26 0 0.013744748 0.018299395 0.026644990770373685 1 -27 0 -0.5081046 0.0040456867 0.0058485307716924636 0 -29 0 -0.4371322 0.0049725696 0.0071917972788715654 0 -30 0 -0.33233207 0.006740546 0.0097574744977965283 0 -33 0 -0.71671814 0.0022043097 0.0031836568219844757 0 -34 0 -0.5343334 0.0037485536 0.005418181176808378 0 -36 1 3.160315 0.9945385 0.0079008942731758486 1 -38 1 2.4890094 0.96247697 0.055176076213957793 1 -39 1 1.6273795 0.67455816 0.56798525229849517 1 -42 1 2.8287919 0.9857495 0.020707046051323227 1 -43 1 0.7827556 0.14967467 2.7400980159955224 1 -47 0 -0.68413067 0.0024238066 0.0035010583956992087 0 -49 1 2.631242 0.9749088 0.036660787288181659 1 -53 1 2.088816 0.88856035 0.17045832146626333 1 -55 1 2.5466716 0.9681056 0.046763649918612135 1 -57 1 -0.008274963 0.017179515 5.8631668640081562 0 -58 1 1.67213 0.7025627 0.50930113190080484 1 -59 1 1.9257294 0.8320127 0.26532252231703396 1 -61 0 -0.3415896 0.0065619755 0.0094981268874324339 0 -62 1 2.6720922 0.97766685 0.032585151920949615 1 -65 1 1.5094963 0.59500456 0.74902737337203174 1 -67 1 1.9699949 0.84931016 0.23563658690189204 1 -75 0 -0.34618023 0.006475177 0.009372081728321581 0 -78 0 0.081366315 0.022204941 0.032395979962873106 1 -80 0 -0.5195276 0.0039135017 0.0056570662610288744 0 -81 0 -0.44965026 0.004794966 0.0069343118297413089 0 -83 0 -0.9477993 0.0011239104 0.0016223718613606979 0 -84 1 2.3634307 0.94674337 0.078954683236577061 1 -85 1 2.0408435 0.8739182 0.19442988729177321 1 -86 1 1.6837547 0.70960605 0.49490978204591057 1 -87 1 2.3847747 0.9497994 0.074305222035416993 1 -89 0 -0.8817291 0.0013627145 0.0019673221341260571 0 -94 0 -0.8288115 0.0015900306 0.002295754930499961 0 -101 1 1.2676952 0.4203619 1.2502961584524357 1 -103 1 0.0039982027 0.017795129 5.8123738189854279 1 -107 1 2.1038253 0.8928264 0.16354844300471233 1 -110 0 0.3695603 0.050042156 0.074064602274564378 1 -114 0 0.6968278 0.120464765 0.18518671956449051 1 -116 0 1.2443513 0.403851 0.74625514145698379 1 -118 0 -0.392108 0.0056671966 0.0081992919182101547 0 -119 0 0.071936555 0.02161497 0.031525766075604846 1 -124 1 2.3050582 0.93746644 0.093161045973030321 1 -126 1 2.722806 0.9806811 0.028143989466295965 1 -127 0 -0.79000276 0.0017804598 0.0025709499022710065 0 -130 0 -0.2646816 0.008200438 0.011879506314543299 0 -134 0 -0.96161973 0.0010795102 0.0015582452327672419 0 -135 0 0.16936706 0.028524533 0.041750531357232966 1 -136 0 -0.57446826 0.0033354373 0.0048200617941456031 0 +24 0 -2.2235482 0.00037039092 0.00053446012476124989 0 +26 0 -0.34385562 0.035934366 0.052796726221873917 0 +27 0 -1.206412 0.0044720713 0.0064663047831698735 0 +29 0 -0.6809915 0.016039982 0.023328400905241749 0 +30 0 -0.7033642 0.015196399 0.022092057498527284 0 +33 0 -2.687782 0.0001186275 0.00017115345316149846 0 +34 0 -1.0021595 0.0073595834 0.010656897912599926 0 +36 1 2.8163698 0.9886031 0.016536642480113755 1 +38 1 2.3243263 0.96288496 0.054564647659344719 1 +39 1 1.3807555 0.71934545 0.47524333410656777 1 +42 1 2.6892452 0.9844971 0.022541181956994991 1 +43 1 -0.39085677 0.032146886 4.959177196631722 0 +47 0 -1.981279 0.00067086925 0.0009681845428198021 0 +49 1 2.497702 0.9754269 0.035894317595296238 1 +53 1 1.8052199 0.8789453 0.18615472564869165 1 +55 1 2.370473 0.96672523 0.048822200359144009 1 +57 1 -7.2032223 1.8347986E-09 29.021731105823392 0 +58 1 1.4358134 0.7457895 0.42315955543578576 1 +59 1 1.4597285 0.75675076 0.40210987124953135 1 +61 0 -0.61381143 0.018859481 0.027468321426238659 0 +62 1 2.5531964 0.9784872 0.031375125867803812 1 +65 1 0.6378821 0.2929411 1.7713175078475203 1 +67 1 1.7717317 0.86992884 0.20103070652644711 1 +75 0 -0.74836457 0.013629773 0.019798840944884856 0 +78 0 -0.40451366 0.0311207 0.04561114569332423 0 +80 0 -3.5225036 1.5308642E-05 2.2085871100526658E-05 0 +81 0 -1.0771459 0.006130582 0.0088717829887540803 0 +83 0 -3.907708 5.9504896E-06 8.5847673705082464E-06 0 +84 1 2.1456301 0.94361675 0.083727070296234579 1 +85 1 1.9041718 0.9024916 0.14801454424794211 1 +86 1 1.4982431 0.77371854 0.37011925811953716 1 +87 1 2.204877 0.9508691 0.072681371808393944 1 +89 0 -2.7511554 0.00010154914 0.00014651188484762575 0 +94 0 -2.5031815 0.00018656286 0.00026917842059635231 0 +101 1 1.1091447 0.56830096 0.8152729362147233 1 +103 1 -7.1133304 2.2874795E-09 28.703594015044942 0 +107 1 1.8841872 0.8980915 0.1550656647180832 1 +110 0 -0.039298795 0.0729421 0.10926864949605086 0 +114 0 0.3733886 0.177997 0.28278442350783617 1 +116 0 0.6278228 0.28785613 0.48975936916879015 1 +118 0 -0.75135326 0.013531557 0.019655195246958149 0 +119 0 -0.40421265 0.031142974 0.045644312108186307 0 +124 1 1.9284774 0.9076142 0.13984895943337902 1 +126 1 2.5906727 0.9803396 0.028646513787097472 1 +127 0 -2.782698 9.398853E-05 0.00013560315668595824 0 +130 0 -0.985148 0.0076708007 0.011109289337923798 0 +134 0 -2.8460355 8.0464124E-05 0.00011608986331816383 0 +135 0 -0.5472505 0.022130538 0.032286205144039654 0 +136 0 -1.1420736 0.0052326466 0.0075689333865708005 0 139 0 ? ? ? 0 -140 0 -0.48607594 0.0043132827 0.0062362104745317592 0 -142 1 2.284 0.9337636 0.098870708488602327 1 -143 0 0.4042234 0.055078276 0.081733270691634122 1 -146 1 1.4803417 0.57433766 0.80002892765984768 1 -148 0 -1.5704706 0.00018263912 0.00026351662273355547 0 -149 1 2.9833307 0.9908772 0.013221804298425196 1 -153 0 0.27521652 0.038456753 0.056576348108561957 1 -155 1 2.164162 0.90855396 0.13835589794076927 1 -157 0 -0.8288115 0.0015900306 0.002295754930499961 0 +140 0 -0.87579167 0.010007432 0.01451039975689827 0 +142 1 2.0708084 0.933018 0.10002313630026556 1 +143 0 0.24293041 0.13587205 0.21068315107693769 1 +146 1 1.3087332 0.68233806 0.55144140676533016 1 +148 0 -6.604278 7.974254E-09 1.1504416956937228E-08 0 +149 1 2.8255935 0.98885524 0.016168752752262736 1 +153 0 -0.13982724 0.057923757 0.086084271131345216 0 +155 1 2.092022 0.9361979 0.095114601516294472 1 +157 0 -2.5031815 0.00018656286 0.00026917842059635231 0 158 0 ? ? ? 0 -159 1 3.5016708 0.9979771 0.0029214154889689765 1 -160 1 3.1946273 0.9950565 0.0071496352202290179 1 -162 0 -0.79000276 0.0017804598 0.0025709499022710065 0 -163 0 0.027284654 0.019023262 0.027709168757291309 1 -165 0 -0.080960155 0.013940394 0.020253236529536374 0 -166 1 2.4831634 0.96185565 0.056107696453336915 1 -168 0 -0.79000276 0.0017804598 0.0025709499022710065 0 -170 0 -0.48607594 0.0043132827 0.0062362104745317592 0 -172 0 -0.68413067 0.0024238066 0.0035010583956992087 0 -175 1 2.515512 0.96517366 0.051139548003930106 1 -178 0 -0.7354426 0.002087281 0.0030144571763673074 0 -182 0 -0.5865108 0.0032205714 0.0046538004195265494 0 -184 1 2.7342782 0.9813056 0.027225602876288377 1 -185 0 -0.51000273 0.0040234183 0.0058162741370629319 0 -186 1 2.130156 0.8999635 0.15206160695130336 1 -190 1 3.6633766 0.9987374 0.0018227051674108922 1 -193 0 -0.8497671 0.0014958045 0.0021596053416366974 0 -194 0 -0.79000276 0.0017804598 0.0025709499022710065 0 -195 0 -0.7354426 0.002087281 0.0030144571763673074 0 -197 0 -0.20865189 0.009643845 0.01398065022571535 0 -200 1 3.2660077 0.99598277 0.0058073158500071519 1 -203 0 -0.66704065 0.0025475002 0.0036799552259266476 0 -208 0 -0.5381637 0.0037070205 0.0053580374793541627 0 -213 1 3.839589 0.9992448 0.0010899216944802554 1 -214 1 4.047656 0.9995885 0.00059380629791341746 1 -215 1 2.8016074 0.9845905 0.022404318151562994 1 -217 0 -0.8497671 0.0014958045 0.0021596053416366974 0 -220 0 -0.7571008 0.001959629 0.0028299205868849656 0 -221 1 3.1183271 0.99383056 0.0089281874876965435 1 -222 1 0.9670042 0.23161604 2.1101929582963121 1 -224 1 3.3201737 0.9965684 0.0049592926536500131 1 -225 0 -0.68413067 0.0024238066 0.0035010583956992087 0 -227 1 3.0194664 0.99178314 0.011903391647371362 1 -229 1 3.6313276 0.9986137 0.0020013738046940669 1 -230 1 2.5642788 0.9696553 0.044456152684478861 1 -231 1 2.8629768 0.9870857 0.018752748150742782 1 -232 0 1.6048453 0.6599523 1.556190888567915 1 -234 0 0.5285514 0.07731862 0.11609555767347124 1 +159 1 3.3141904 0.99661213 0.0048959590695084624 1 +160 1 3.0384002 0.99335766 0.0096148413053643873 1 +162 0 -2.782698 9.398853E-05 0.00013560315668595824 0 +163 0 -0.16804563 0.05425993 0.080484371122169787 0 +165 0 -0.71078014 0.014926531 0.0216967673728562 0 +166 1 2.3001277 0.9607042 0.057835788941002261 1 +168 0 -2.782698 9.398853E-05 0.00013560315668595824 0 +170 0 -0.87579167 0.010007432 0.01451039975689827 0 +172 0 -1.981279 0.00067086925 0.0009681845428198021 0 +175 1 2.3206725 0.9625633 0.055046712228318653 1 +178 0 -3.0617628 4.7400794E-05 6.8386511016753427E-05 0 +182 0 -3.6175222 1.2125714E-05 1.7493814123837223E-05 0 +184 1 2.6413684 0.98259866 0.025325819755526953 1 +185 0 -0.8295902 0.011194982 0.016242029686661073 0 +186 1 1.8576343 0.89197344 0.16492734942412537 1 +190 1 3.3874273 0.9971677 0.0040919333192415332 1 +193 0 -2.2235482 0.00037039092 0.00053446012476124989 0 +194 0 -2.782698 9.398853E-05 0.00013560315668595824 0 +195 0 -3.0617628 4.7400794E-05 6.8386511016753427E-05 0 +197 0 -1.1602051 0.0050061448 0.0072404788133332287 0 +200 1 3.0746958 0.99392 0.008798319102579279 1 +203 0 -3.340107 2.394717E-05 3.4548877251977815E-05 0 +208 0 -1.917244 0.0007848919 0.0011328042604764636 0 +213 1 3.594258 0.9982928 0.0024650683199034479 1 +214 1 3.7810779 0.9989198 0.0015592631121796418 1 +215 1 2.5531583 0.97848517 0.031378113855905755 1 +217 0 -2.2235482 0.00037039092 0.00053446012476124989 0 +220 0 -2.4086835 0.00023522312 0.0003393951529091138 0 +221 1 2.8388205 0.98920727 0.015655255559260929 1 +222 1 0.7111739 0.3315133 1.5928614063341144 1 +224 1 3.113253 0.9944657 0.0080064701273409138 1 +225 0 -1.981279 0.00067086925 0.0009681845428198021 0 +227 1 2.7846591 0.98769253 0.017866088883108887 1 +229 1 3.3525598 0.9969156 0.0044567559033219777 1 +230 1 2.3219533 0.96267635 0.054877252237347518 1 +231 1 2.621455 0.9817434 0.026582107683329387 1 +232 0 0.9736352 0.4856292 0.95911935198904652 1 +234 0 0.049498107 0.08911245 0.13465513705549487 1 235 0 ? ? ? 0 -236 1 3.1184926 0.99383354 0.008923861237336652 1 -238 1 3.2699842 0.99602896 0.0057404053229708823 1 -243 0 0.41025028 0.05600128 0.083143190738845754 1 -245 0 -0.004266075 0.017378261 0.025291938487990686 0 -251 1 3.0358403 0.9921637 0.011349894342852419 1 -253 1 2.8287919 0.9857495 0.020707046051323227 1 -255 1 2.2902777 0.9348883 0.097134087031752178 1 -256 0 -0.48607594 0.0043132827 0.0062362104745317592 0 -261 1 3.03942 0.99224454 0.011232373931618893 1 -263 1 2.9294598 0.9893399 0.01546185031359365 1 -264 1 2.094007 0.8900523 0.16803795259780735 1 -265 0 -0.076622814 0.014115545 0.020509521224697652 0 -266 1 3.1994135 0.9951248 0.0070506029786410011 1 -270 1 2.6933858 0.9789848 0.030641674137859447 1 -273 1 1.3901646 0.5090698 0.97406461195856153 1 -274 0 -0.6331637 0.0028116107 0.0040620098868226762 0 -281 0 -0.71671814 0.0022043097 0.0031836568219844757 0 -282 1 1.8982561 0.8205011 0.28542284617889779 1 -286 1 3.545079 0.99821746 0.0025739513416233248 1 -287 0 -0.96161973 0.0010795102 0.0015582452327672419 0 -289 1 2.5929847 0.9720266 0.040932320424446333 1 +236 1 3.0514817 0.99356604 0.009312237158435363 1 +238 1 3.1560044 0.995014 0.0072112529512283049 1 +243 0 0.046878263 0.08859216 0.13383130898338974 1 +245 0 -0.793629 0.012214831 0.017730787318238534 0 +251 1 2.918702 0.9911108 0.012881741690830003 1 +253 1 2.6892452 0.9844971 0.022541181956994991 1 +255 1 2.1789172 0.94780713 0.077334576151134218 1 +256 0 -0.87579167 0.010007432 0.01451039975689827 0 +261 1 2.8177457 0.9886411 0.016481235591725343 1 +263 1 2.7477593 0.98654217 0.019547381151436293 1 +264 1 1.7881138 0.8744089 0.19362000839621532 1 +265 0 -0.87861484 0.00993905 0.014410752401667519 0 +266 1 3.1317565 0.99471 0.0076520736422275521 1 +270 1 2.552394 0.9784457 0.031436292975207218 1 +273 1 0.91059536 0.44716215 1.1611300140178031 1 +274 0 -2.9622543 6.050539E-05 8.7293467188612008E-05 0 +281 0 -2.687782 0.0001186275 0.00017115345316149846 0 +282 1 1.7275698 0.85717165 0.22234395170689594 1 +286 1 3.4030685 0.99727404 0.003938097264721962 1 +287 0 -2.8460355 8.0464124E-05 0.00011608986331816383 0 +289 1 2.4191728 0.9703611 0.043406358129627787 1 292 1 ? ? ? 0 294 0 ? ? ? 0 -295 1 2.537888 0.96730417 0.047958476396403632 1 -298 0 -0.52101225 0.0038966408 0.0056326457651658703 0 -302 1 3.6927981 0.9988412 0.0016727267010132856 1 -305 1 3.1617885 0.9945618 0.0078670874222826776 1 -306 0 -0.8497671 0.0014958045 0.0021596053416366974 0 -307 0 -0.8497671 0.0014958045 0.0021596053416366974 0 -310 0 -0.99169445 0.0009888523 0.0014273182000734896 0 -313 0 -0.538535 0.003703019 0.0053522431718029638 0 +295 1 2.3347461 0.96378773 0.053212653921867824 1 +298 0 -1.9893144 0.00065778324 0.00094929286580709086 0 +302 1 3.5252311 0.99797845 0.0029194336807990672 1 +305 1 3.0512433 0.99356234 0.0093176031551556217 1 +306 0 -2.2235482 0.00037039092 0.00053446012476124989 0 +307 0 -2.2235482 0.00037039092 0.00053446012476124989 0 +310 0 -2.566779 0.0001596179 0.00023029834224341892 0 +313 0 -0.74154985 0.013856358 0.020130289400489415 0 315 0 ? ? ? 0 -318 0 -2.0152178 4.9855396E-05 7.1927926095832038E-05 0 -320 1 2.5674763 0.96992874 0.044049335346043113 1 -322 0 -0.79000276 0.0017804598 0.0025709499022710065 0 -324 0 -0.8497671 0.0014958045 0.0021596053416366974 0 -325 0 0.11218868 0.02424517 0.035409397140683196 1 -326 1 2.4516737 0.95833546 0.061397344294885685 1 -330 1 2.7521803 0.98224074 0.025851438200705536 1 -334 1 2.7059832 0.9797283 0.029546408200462206 1 -335 0 -0.538535 0.003703019 0.0053522431718029638 0 -337 0 -0.8497671 0.0014958045 0.0021596053416366974 0 -339 1 2.6817114 0.97827196 0.031692502359053562 1 -340 1 2.8886113 0.98800576 0.017408646144306816 1 -341 0 -0.8497671 0.0014958045 0.0021596053416366974 0 -342 0 -0.5212217 0.003894268 0.0056292091843818106 0 -345 0 -0.538535 0.003703019 0.0053522431718029638 0 -351 0 -0.8288115 0.0015900306 0.002295754930499961 0 -356 1 1.035413 0.2690443 1.8940843000125087 1 -357 1 3.1527798 0.99441767 0.0080761665299991219 1 -359 1 2.4969456 0.9633049 0.053935624616307951 1 -362 0 0.15804514 0.027622651 0.040411808381539606 1 -363 0 1.1272937 0.3249282 0.56688712866456559 1 -364 0 -0.8288115 0.0015900306 0.002295754930499961 0 -365 0 -0.6645783 0.0025658335 0.0037064724080739684 0 -367 1 3.155404 0.99446005 0.0080146847888438594 1 -369 0 -0.2931926 0.007550395 0.010934248649784031 0 -372 0 -0.46696994 0.004559601 0.0065931565582987494 0 -374 0 -0.5343334 0.0037485536 0.005418181176808378 0 -375 0 -0.538535 0.003703019 0.0053522431718029638 0 -380 0 -0.538535 0.003703019 0.0053522431718029638 0 -382 0 0.03387748 0.019385818 0.028242468907723462 1 -385 0 0.18651512 0.029945167 0.043861795653076387 1 -386 1 2.727801 0.98095554 0.027740342784481151 1 -390 0 -0.609339 0.0030135398 0.0043541830153293135 0 -393 0 -0.28000757 0.007844398 0.011361695121371914 0 -394 0 -0.2055268 0.00973138 0.014108171929768289 0 -397 0 -0.43539375 0.004997747 0.0072283025366715846 0 -400 1 2.6732848 0.9777428 0.032473100772100201 1 -401 0 -0.48607594 0.0043132827 0.0062362104745317592 0 -402 0 0.5095914 0.073460676 0.11007588640708431 1 -403 0 0.4241748 0.05818972 0.086491624802327319 1 -405 0 -0.68413067 0.0024238066 0.0035010583956992087 0 -407 0 -0.68413067 0.0024238066 0.0035010583956992087 0 -408 0 0.42441255 0.058227774 0.086549918067053574 1 -410 0 -0.68413067 0.0024238066 0.0035010583956992087 0 +318 0 -3.1379292 3.9322684E-05 5.6731756246082198E-05 0 +320 1 2.47477 0.9740418 0.037944380490835941 1 +322 0 -2.782698 9.398853E-05 0.00013560315668595824 0 +324 0 -2.2235482 0.00037039092 0.00053446012476124989 0 +325 0 -0.3330837 0.036861118 0.054184249022080244 0 +326 1 2.2742786 0.9582392 0.061542265329069173 1 +330 1 2.5871398 0.98017186 0.028893367322023938 1 +334 1 2.6192565 0.98164654 0.026724449152624889 1 +335 0 -0.74154985 0.013856358 0.020130289400489415 0 +337 0 -2.2235482 0.00037039092 0.00053446012476124989 0 +339 1 2.581934 0.9799221 0.029261006486360441 1 +340 1 2.7146688 0.98542076 0.021188224341192583 1 +341 0 -2.2235482 0.00037039092 0.00053446012476124989 0 +342 0 -0.8119284 0.01168489 0.016956997114979482 0 +345 0 -0.74154985 0.013856358 0.020130289400489415 0 +351 0 -2.5031815 0.00018656286 0.00026917842059635231 0 +356 1 0.6004559 0.2742922 1.8662144935604821 1 +357 1 3.037542 0.9933437 0.0096350979684566521 1 +359 1 2.3367383 0.9639579 0.052957946838511155 1 +362 0 -0.37844574 0.033107772 0.048573002984905231 0 +363 0 0.9072117 0.44511113 0.84972921812670454 1 +364 0 -2.5031815 0.00018656286 0.00026917842059635231 0 +365 0 -2.2609096 0.000337964 0.00048766141349285949 0 +367 1 2.8964586 0.9906169 0.013600835570715222 1 +369 0 -0.52511513 0.023336649 0.034066733772181217 0 +372 0 -1.0531003 0.0065006623 0.0094090891675049134 0 +374 0 -1.0021595 0.0073595834 0.010656897912599926 0 +375 0 -0.74154985 0.013856358 0.020130289400489415 0 +380 0 -0.74154985 0.013856358 0.020130289400489415 0 +382 0 -0.51982236 0.023634415 0.03450665126135876 0 +385 0 -0.25568157 0.04422791 0.065261454517767589 0 +386 1 2.6102216 0.9812429 0.027317792062589313 1 +390 0 -2.1669154 0.00042557062 0.00061409929686916261 0 +393 0 -0.44209728 0.028457755 0.041651366183955311 0 +394 0 -0.49994117 0.024786381 0.036209821417884709 0 +397 0 -0.9341007 0.008685227 0.012584863928127092 0 +400 1 2.54729 0.97818005 0.031828052449825284 1 +401 0 -0.87579167 0.010007432 0.01451039975689827 0 +402 0 -0.018790785 0.076418 0.11468803473258497 0 +403 0 0.13496335 0.107660525 0.16433543140072923 1 +405 0 -1.981279 0.00067086925 0.0009681845428198021 0 +407 0 -1.981279 0.00067086925 0.0009681845428198021 0 +408 0 0.06905651 0.09308453 0.14096000423614075 1 +410 0 -1.981279 0.00067086925 0.0009681845428198021 0 411 0 ? ? ? 0 -412 1 3.1056695 0.9935998 0.009263251724017851 1 -417 0 -0.68413067 0.0024238066 0.0035010583956992087 0 -420 0 0.6235068 0.09956116 0.15129981046427482 1 -421 1 3.4023418 0.99729836 0.003902917332663001 1 -424 0 -0.48607594 0.0043132827 0.0062362104745317592 0 -425 1 3.771371 0.9990786 0.0013299527219586516 1 -426 0 0.77619374 0.14725265 0.2298097267894085 1 -427 1 2.453265 0.95852053 0.061118759917589138 1 -431 0 -0.8456709 0.0015137739 0.0021855687803360216 0 -432 0 -0.34406832 0.006514966 0.009429860244496149 0 -433 0 0.1209803 0.024859898 0.036318583471795833 1 -435 1 3.1685188 0.99466705 0.0077144044554509931 1 -437 0 -0.43539375 0.004997747 0.0072283025366715846 0 -438 0 -0.018248942 0.01669468 0.024288646313225117 0 -443 0 -0.24981476 0.008561113 0.012404248324275951 0 -444 0 -0.09983369 0.013202864 0.019174566032128752 0 -445 0 -0.5212217 0.003894268 0.0056292091843818106 0 -446 0 -0.538535 0.003703019 0.0053522431718029638 0 -447 0 -0.37125614 0.006020797 0.008712428182113121 0 -448 0 -0.28000757 0.007844398 0.011361695121371914 0 -458 0 -0.26429355 0.008209658 0.011892918142384705 0 -459 0 -0.15420155 0.011286834 0.016376051533720463 0 -460 0 -0.13420394 0.011957337 0.01735475737350195 0 -461 0 0.622342 0.0992567 0.15081208222655509 1 -462 0 -0.03996976 0.015684903 0.022807872219889662 0 -463 0 -0.30480245 0.0073005855 0.010571153437721885 0 -468 0 -0.43539375 0.004997747 0.0072283025366715846 0 -469 0 -0.62637657 0.0028677206 0.0041431898228888502 0 -470 0 -0.33233207 0.006740546 0.0097574744977965283 0 -471 0 -0.03996976 0.015684903 0.022807872219889662 0 -472 0 -0.018656634 0.01667515 0.024259992633796853 0 -473 0 -0.43539375 0.004997747 0.0072283025366715846 0 -475 0 -0.48607594 0.0043132827 0.0062362104745317592 0 -476 0 -0.34099272 0.006573346 0.0095146395257793542 0 -477 0 -0.43539375 0.004997747 0.0072283025366715846 0 -478 0 0.033098403 0.019342626 0.028178924263703012 1 -479 1 2.787153 0.9839369 0.023362287981379421 1 -481 0 0.6781833 0.11481523 0.1759494619793455 1 -485 0 0.26776877 0.037660692 0.055382436003477434 1 -486 0 -0.33233207 0.006740546 0.0097574744977965283 0 -488 1 1.6145483 0.6662808 0.58579776123067451 1 -490 0 -0.538535 0.003703019 0.0053522431718029638 0 -491 1 2.72395 0.9807443 0.028051046034553873 1 -494 0 1.2574471 0.41308928 0.76878702411508326 1 -496 0 -0.28000757 0.007844398 0.011361695121371914 0 -498 0 -0.57446826 0.0033354373 0.0048200617941456031 0 -499 0 -0.57446826 0.0033354373 0.0048200617941456031 0 -500 0 -0.5865108 0.0032205714 0.0046538004195265494 0 -503 0 -0.7354426 0.002087281 0.0030144571763673074 0 -505 0 0.15788618 0.027610188 0.04039331748034708 1 -506 1 2.9675121 0.9904501 0.01384382555563791 1 -508 0 -0.37125614 0.006020797 0.008712428182113121 0 -509 0 -0.5212217 0.003894268 0.0056292091843818106 0 -511 0 -0.5081046 0.0040456867 0.0058485307716924636 0 -512 0 -0.37125614 0.006020797 0.008712428182113121 0 -515 1 3.0504782 0.9924891 0.010876837649353148 1 -516 0 -0.28000757 0.007844398 0.011361695121371914 0 -518 0 -0.30709174 0.007252304 0.010500987216249735 0 -524 0 -0.6272879 0.002860122 0.004132195786456465 0 -525 0 -0.39318767 0.005649461 0.008173559192018183 0 -526 0 -0.43539375 0.004997747 0.0072283025366715846 0 -530 1 2.6465328 0.9759781 0.035079354210466686 1 -536 0 -0.66704065 0.0025475002 0.0036799552259266476 0 -537 0 -0.5328945 0.0037642755 0.0054409485922593667 0 -542 0 0.3557345 0.04815768 0.071205496270431268 1 -543 0 -0.57446826 0.0033354373 0.0048200617941456031 0 -545 0 -0.5081046 0.0040456867 0.0058485307716924636 0 -550 0 -0.6272879 0.002860122 0.004132195786456465 0 -551 0 -0.8497671 0.0014958045 0.0021596053416366974 0 -552 0 -0.041153874 0.015631616 0.022729773078318208 0 -553 0 1.2232385 0.3891023 0.71099730961492824 1 -554 0 -0.48607594 0.0043132827 0.0062362104745317592 0 -555 0 0.88857573 0.19338019 0.31003926059814191 1 -556 0 0.38323748 0.05197496 0.077002929138665591 1 -562 0 -0.8497671 0.0014958045 0.0021596053416366974 0 -564 0 -0.56291914 0.0034494305 0.004985079006480899 0 -567 0 0.2355448 0.034395326 0.050495435320072148 1 -568 1 2.2316751 0.92366534 0.11455785468815727 1 -570 1 2.9043825 0.98853934 0.016629716925098629 1 -571 1 3.3564863 0.99691254 0.0044611550363844003 1 -572 0 -0.6272879 0.002860122 0.004132195786456465 0 -573 0 -0.68413067 0.0024238066 0.0035010583956992087 0 -574 1 2.8435903 0.98634386 0.019837406157358722 1 -575 0 -0.5328945 0.0037642755 0.0054409485922593667 0 -576 0 -0.5081046 0.0040456867 0.0058485307716924636 0 -579 0 -0.8497671 0.0014958045 0.0021596053416366974 0 -580 0 -0.38784665 0.00573774 0.0083016483516430081 0 -583 0 -0.48607594 0.0043132827 0.0062362104745317592 0 -585 0 -0.538535 0.003703019 0.0053522431718029638 0 -587 0 -0.34406832 0.006514966 0.009429860244496149 0 -588 1 2.4578753 0.95905244 0.06031838713618997 1 -589 0 -0.37125614 0.006020797 0.008712428182113121 0 -591 1 2.2657692 0.9303946 0.10408538709787052 1 -592 1 2.636997 0.9753166 0.036057506598704957 1 -595 0 -0.5081046 0.0040456867 0.0058485307716924636 0 -596 0 -0.46696994 0.004559601 0.0065931565582987494 0 -597 0 -0.39588702 0.00560536 0.008109574950900306 0 -598 0 -0.6272879 0.002860122 0.004132195786456465 0 -599 0 0.5072477 0.07299628 0.10935296828466248 1 -601 0 -0.40043423 0.005531842 0.0080029167602950237 0 -603 1 2.0076923 0.86286175 0.21279866519937116 1 -605 1 3.1730316 0.9947365 0.0076136909583056547 1 -608 1 2.9642394 0.9903593 0.013976059166149711 1 -610 1 3.0031703 0.9913862 0.012480954269021304 1 -611 1 2.5794322 0.9709304 0.042560217127043151 1 -615 0 -0.3546906 0.006317271 0.0091428047172382593 0 -616 0 -0.6272879 0.002860122 0.004132195786456465 0 -620 0 -0.6272879 0.002860122 0.004132195786456465 0 -623 0 -0.538535 0.003703019 0.0053522431718029638 0 -625 0 0.433428 0.05968811 0.088788734201066505 1 -626 1 2.1038232 0.8928258 0.1635494061410647 1 -628 0 -0.5212217 0.003894268 0.0056292091843818106 0 -630 0 0.33890998 0.045955323 0.067871266565750571 1 -631 0 -0.5081046 0.0040456867 0.0058485307716924636 0 -632 0 -0.538535 0.003703019 0.0053522431718029638 0 -635 0 -0.22692177 0.009147483 0.013257758684029619 0 -636 1 3.1565762 0.9944789 0.007987360411549644 1 -637 0 0.6592889 0.109326616 0.16703161215114079 1 -640 0 -0.22026803 0.009325254 0.013516618739641769 0 -643 0 -0.538535 0.003703019 0.0053522431718029638 0 -646 0 -0.20056786 0.0098719 0.014312905895707762 0 -647 0 -0.41501325 0.0053025368 0.0076702975350839342 0 -648 1 2.6994166 0.979344 0.030112376168537937 1 -650 0 0.35254174 0.047732178 0.070560711526742523 1 -651 0 -0.1919981 0.010119489 0.014673708240985506 0 -655 0 -0.6272879 0.002860122 0.004132195786456465 0 -658 1 3.0571246 0.9926324 0.010668565109978326 1 -659 0 -0.538535 0.003703019 0.0053522431718029638 0 -662 0 -0.5572568 0.0035067294 0.0050680323969575997 0 -663 0 -0.5572568 0.0035067294 0.0050680323969575997 0 -664 0 -0.40340298 0.0054843617 0.0079340378878566178 0 -666 0 0.1943484 0.030616714 0.04486088649315996 1 -667 0 -0.79000276 0.0017804598 0.0025709499022710065 0 -669 1 2.8635 0.98710513 0.018724348492869053 1 -671 0 -0.20638002 0.009707403 0.01407324106877276 0 -672 0 -0.8288115 0.0015900306 0.002295754930499961 0 -673 0 0.26565865 0.037438042 0.055048688783445286 1 -674 0 -0.68413067 0.0024238066 0.0035010583956992087 0 -675 0 -0.13901675 0.011792458 0.017114027918625194 0 -676 0 -0.62637657 0.0028677206 0.0041431898228888502 0 -677 0 -0.37125614 0.006020797 0.008712428182113121 0 -684 0 -0.538535 0.003703019 0.0053522431718029638 0 -686 0 -0.538535 0.003703019 0.0053522431718029638 0 -687 0 -0.35511437 0.006309509 0.0091315352137533584 0 -690 0 -0.41501325 0.0053025368 0.0076702975350839342 0 -695 0 -0.5212217 0.003894268 0.0056292091843818106 0 +412 1 2.8262787 0.9888738 0.016141708296823532 1 +417 0 -1.981279 0.00067086925 0.0009681845428198021 0 +420 0 0.20248847 0.12463884 0.19204972315723823 1 +421 1 3.2064 0.9955912 0.0063745827725976674 1 +424 0 -0.87579167 0.010007432 0.01451039975689827 0 +425 1 3.4421244 0.9975225 0.0035787504906085615 1 +426 0 0.36153084 0.17378072 0.27540337640340673 1 +427 1 2.169831 0.9466934 0.079030799567607876 1 +431 0 -4.4777765 1.4696584E-06 2.1202704439656894E-06 0 +432 0 -1.0489565 0.0065666446 0.0095049076092876272 0 +433 0 -0.31749022 0.038243502 0.056256423492893781 0 +435 1 3.0718942 0.99387836 0.0088587959967189427 1 +437 0 -0.9341007 0.008685227 0.012584863928127092 0 +438 0 -0.61796576 0.018671831 0.027192421797524614 0 +443 0 -0.47722003 0.026169965 0.038258097859764466 0 +444 0 -1.1696978 0.0048914794 0.0070742289177469237 0 +445 0 -0.8119284 0.01168489 0.016956997114979482 0 +446 0 -0.74154985 0.013856358 0.020130289400489415 0 +447 0 -0.98767114 0.0076238294 0.011041001927285387 0 +448 0 -0.44209728 0.028457755 0.041651366183955311 0 +458 0 -0.8921775 0.0096169375 0.013941453032426377 0 +459 0 -0.78838825 0.012370926 0.017958787947952616 0 +460 0 -0.83895427 0.010943532 0.015875204042194665 0 +461 0 0.39383703 0.18545538 0.29593436778609289 1 +462 0 -0.86123663 0.0103674345 0.015035119479152098 0 +463 0 -0.8140986 0.011623569 0.016867486328837465 0 +468 0 -0.9341007 0.008685227 0.012584863928127092 0 +469 0 -0.8865646 0.00974897 0.014133798382209318 0 +470 0 -0.7033642 0.015196399 0.022092057498527284 0 +471 0 -0.86123663 0.0103674345 0.015035119479152098 0 +472 0 -0.6326933 0.018021226 0.026236254536640333 0 +473 0 -0.9341007 0.008685227 0.012584863928127092 0 +475 0 -0.87579167 0.010007432 0.01451039975689827 0 +476 0 -0.8555222 0.010512251 0.015246250326461824 0 +477 0 -0.9341007 0.008685227 0.012584863928127092 0 +478 0 -0.5253376 0.023324216 0.034048368038637533 0 +479 1 2.5875406 0.98019093 0.028865293718249165 1 +481 0 0.20480001 0.12525883 0.19307190373867764 1 +485 0 0.06191207 0.09161549 0.13862499080702864 1 +486 0 -0.7033642 0.015196399 0.022092057498527284 0 +488 1 1.4248846 0.7406733 0.43309075640568206 1 +490 0 -0.74154985 0.013856358 0.020130289400489415 0 +491 1 2.588621 0.9802423 0.028789673165762576 1 +494 0 0.8140745 0.38961735 0.71221414913157211 1 +496 0 -0.44209728 0.028457755 0.041651366183955311 0 +498 0 -1.1420736 0.0052326466 0.0075689333865708005 0 +499 0 -1.1420736 0.0052326466 0.0075689333865708005 0 +500 0 -3.6175222 1.2125714E-05 1.7493814123837223E-05 0 +503 0 -3.0617628 4.7400794E-05 6.8386511016753427E-05 0 +505 0 -0.17351717 0.053575255 0.079440300314024756 0 +506 1 2.7934446 0.98795176 0.017487502233421465 1 +508 0 -0.98767114 0.0076238294 0.011041001927285387 0 +509 0 -0.8119284 0.01168489 0.016956997114979482 0 +511 0 -1.206412 0.0044720713 0.0064663047831698735 0 +512 0 -0.98767114 0.0076238294 0.011041001927285387 0 +515 1 2.8100622 0.98842746 0.01679300314532791 1 +516 0 -0.44209728 0.028457755 0.041651366183955311 0 +518 0 -0.7636737 0.013133994 0.019073882587211946 0 +524 0 -1.0734073 0.006186717 0.008953270531768193 0 +525 0 -0.83473635 0.011056095 0.01603940483861321 0 +526 0 -0.9341007 0.008685227 0.012584863928127092 0 +530 1 2.4531589 0.97266716 0.039981889220973385 1 +536 0 -3.340107 2.394717E-05 3.4548877251977815E-05 0 +537 0 -3.5305064 1.5011042E-05 2.1656518101871654E-05 0 +542 0 -0.111779705 0.061794505 0.092024143537804706 0 +543 0 -1.1420736 0.0052326466 0.0075689333865708005 0 +545 0 -1.206412 0.0044720713 0.0064663047831698735 0 +550 0 -1.0734073 0.006186717 0.008953270531768193 0 +551 0 -2.2235482 0.00037039092 0.00053446012476124989 0 +552 0 -0.67433333 0.016299814 0.023709419626751108 0 +553 0 1.1007556 0.5632451 1.1951042642534468 1 +554 0 -0.87579167 0.010007432 0.01451039975689827 0 +555 0 0.419486 0.19514884 0.31320608355119273 1 +556 0 -0.16065937 0.055197284 0.081914982296408798 0 +562 0 -2.2235482 0.00037039092 0.00053446012476124989 0 +564 0 -1.2197396 0.0043288474 0.006258763082478319 0 +567 0 -0.25599548 0.04419537 0.065212337065460882 0 +568 1 2.1435375 0.943343 0.084145688832568802 1 +570 1 2.768445 0.98719954 0.018586365474098834 1 +571 1 3.0822294 0.9940307 0.008637665331741054 1 +572 0 -1.0734073 0.006186717 0.008953270531768193 0 +573 0 -1.981279 0.00067086925 0.0009681845428198021 0 +574 1 2.571908 0.97943246 0.029982079395301307 1 +575 0 -3.5305064 1.5011042E-05 2.1656518101871654E-05 0 +576 0 -1.206412 0.0044720713 0.0064663047831698735 0 +579 0 -2.2235482 0.00037039092 0.00053446012476124989 0 +580 0 -1.1002328 0.005794982 0.0083847098900298022 0 +583 0 -0.87579167 0.010007432 0.01451039975689827 0 +585 0 -0.74154985 0.013856358 0.020130289400489415 0 +587 0 -1.0489565 0.0065666446 0.0095049076092876272 0 +588 1 2.2345545 0.9541607 0.067695843763695349 1 +589 0 -0.98767114 0.0076238294 0.011041001927285387 0 +591 1 2.0619256 0.9316433 0.1021503904572014 1 +592 1 2.5281107 0.97715294 0.033343704934921212 1 +595 0 -1.206412 0.0044720713 0.0064663047831698735 0 +596 0 -1.0531003 0.0065006623 0.0094090891675049134 0 +597 0 -1.2047621 0.0044901264 0.0064924700138819609 0 +598 0 -1.0734073 0.006186717 0.008953270531768193 0 +599 0 0.07064793 0.09341463 0.14148521018352714 1 +601 0 -0.59420186 0.019770512 0.028808547776574541 0 +603 1 1.4452949 0.75017387 0.41470308911084092 1 +605 1 2.9385715 0.9915301 0.012271496056639943 1 +608 1 2.7334158 0.98606694 0.020242510105247975 1 +610 1 2.855239 0.9896289 0.015040447065429096 1 +611 1 2.2871258 0.95948225 0.059671973538020964 1 +615 0 -0.9541885 0.008271066 0.011982246805936373 0 +616 0 -1.0734073 0.006186717 0.008953270531768193 0 +620 0 -1.0734073 0.006186717 0.008953270531768193 0 +623 0 -0.74154985 0.013856358 0.020130289400489415 0 +625 0 0.0720381 0.093703836 0.14194551677828579 1 +626 1 1.8450893 0.8889721 0.16978994763834329 1 +628 0 -0.8119284 0.01168489 0.016956997114979482 0 +630 0 -0.29814127 0.0400281 0.058935917010753049 0 +631 0 -1.206412 0.0044720713 0.0064663047831698735 0 +632 0 -0.74154985 0.013856358 0.020130289400489415 0 +635 0 -0.6626082 0.01676748 0.024395461094995373 0 +636 1 2.7756643 0.9874214 0.018262192152126763 1 +637 0 0.18570484 0.12021576 0.18477833493845203 1 +640 0 -0.8135531 0.011638952 0.016889941332485767 0 +643 0 -0.74154985 0.013856358 0.020130289400489415 0 +646 0 -0.4364347 0.02884434 0.042225540500301306 0 +647 0 -0.71363914 0.014823757 0.021546256617030777 0 +648 1 2.3701818 0.9667022 0.048856535913411513 1 +650 0 -0.030830527 0.07435937 0.1114759070243895 0 +651 0 -0.53687966 0.022687845 0.033108659980875962 0 +655 0 -1.0734073 0.006186717 0.008953270531768193 0 +658 1 2.925202 0.99125016 0.01267890504973126 1 +659 0 -0.74154985 0.013856358 0.020130289400489415 0 +662 0 -0.83922654 0.010936305 0.015864662267834149 0 +663 0 -0.83922654 0.010936305 0.015864662267834149 0 +664 0 -0.9649343 0.008057617 0.011671770314325982 0 +666 0 -0.39795876 0.031609222 0.046338754728111026 0 +667 0 -2.782698 9.398853E-05 0.00013560315668595824 0 +669 1 2.5719533 0.9794347 0.029978743109534415 1 +671 0 -0.7625153 0.013170878 0.019127804296131281 0 +672 0 -2.5031815 0.00018656286 0.00026917842059635231 0 +673 0 -0.3101213 0.038913965 0.057262510519801864 0 +674 0 -1.981279 0.00067086925 0.0009681845428198021 0 +675 0 -0.67203355 0.01639052 0.023842454083914031 0 +676 0 -0.8865646 0.00974897 0.014133798382209318 0 +677 0 -0.98767114 0.0076238294 0.011041001927285387 0 +684 0 -0.74154985 0.013856358 0.020130289400489415 0 +686 0 -0.74154985 0.013856358 0.020130289400489415 0 +687 0 -0.8366627 0.011004546 0.015964204955816238 0 +690 0 -0.71363914 0.014823757 0.021546256617030777 0 +695 0 -0.8119284 0.01168489 0.016956997114979482 0 diff --git a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer-out.txt b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer-out.txt index 650237ba34..81921808c4 100644 --- a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer-out.txt +++ b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer-out.txt @@ -8,28 +8,28 @@ Confusion table ||====================== PREDICTED || positive | negative | Recall TRUTH ||====================== - positive || 226 | 13 | 0.9456 - negative || 22 | 422 | 0.9505 + positive || 228 | 11 | 0.9540 + negative || 15 | 429 | 0.9662 ||====================== -Precision || 0.9113 | 0.9701 | -OVERALL 0/1 ACCURACY: 0.948755 -LOG LOSS/instance: 0.259132 +Precision || 0.9383 | 0.9750 | +OVERALL 0/1 ACCURACY: 0.961933 +LOG LOSS/instance: 0.188873 Test-set entropy (prior Log-Loss/instance): 0.934003 -LOG-LOSS REDUCTION (RIG): 0.722558 -AUC: 0.971682 +LOG-LOSS REDUCTION (RIG): 0.797781 +AUC: 0.980974 OVERALL RESULTS --------------------------------------- -AUC: 0.971682 (0.0000) -Accuracy: 0.948755 (0.0000) -Positive precision: 0.911290 (0.0000) -Positive recall: 0.945607 (0.0000) -Negative precision: 0.970115 (0.0000) -Negative recall: 0.950450 (0.0000) -Log-loss: 0.259132 (0.0000) -Log-loss reduction: 0.722558 (0.0000) -F1 Score: 0.928131 (0.0000) -AUPRC: 0.961861 (0.0000) +AUC: 0.980974 (0.0000) +Accuracy: 0.961933 (0.0000) +Positive precision: 0.938272 (0.0000) +Positive recall: 0.953975 (0.0000) +Negative precision: 0.975000 (0.0000) +Negative recall: 0.966216 (0.0000) +Log-loss: 0.188873 (0.0000) +Log-loss reduction: 0.797781 (0.0000) +F1 Score: 0.946058 (0.0000) +AUPRC: 0.976643 (0.0000) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer-rp.txt b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer-rp.txt index d831e9838d..384fb65087 100644 --- a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer-rp.txt +++ b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer-rp.txt @@ -1,4 +1,4 @@ LdSvm AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC /bias /iter Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings -0.971682 0.948755 0.91129 0.945607 0.970115 0.95045 0.259132 0.722558 0.928131 0.961861 - 1000 LdSvm %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=LdSvm{iter=1000 bias=-} dout=%Output% data=%Data% out=%Output% seed=1 /bias:-;/iter:1000 +0.980974 0.961933 0.938272 0.953975 0.975 0.966216 0.188873 0.797781 0.946058 0.976643 - 1000 LdSvm %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=LdSvm{iter=1000 bias=-} dout=%Output% data=%Data% out=%Output% seed=1 /bias:-;/iter:1000 diff --git a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer.txt b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer.txt index 758a96884d..8a1cffa058 100644 --- a/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer.txt +++ b/test/BaselineOutput/Common/LdSvm/netcoreapp/LDSVM-nob-TrainTest-breast-cancer.txt @@ -1,700 +1,700 @@ Instance Label Score Probability Log-loss Assigned -0 0 -1.9328692 0.00879855 0.012749796760162687 0 -1 0 1.7016087 0.92350006 3.7083975930171791 1 -2 0 -1.3935351 0.025243035 0.036885536024252437 0 -3 0 2.6855175 0.9883893 6.4284028787783623 1 -4 0 -1.6886145 0.014210899 0.020649064923297558 0 -5 1 3.6676228 0.99833095 0.0024099408096164236 1 -6 0 1.6054935 0.90888304 3.4561364993169899 1 -7 0 -1.533542 0.019235376 0.028021153060704358 0 -8 0 -2.041343 0.007106071 0.010288492033295495 0 -9 0 -1.195721 0.03693594 0.054296330655249615 0 -10 0 -1.0634518 0.047500372 0.070209565626258783 0 -11 0 -1.3867699 0.025575617 0.03737786115106672 0 -12 1 0.56797415 0.55980617 0.83700071192704562 1 -13 0 -0.85706234 0.06987408 0.10450205103638049 0 -14 1 2.2492504 0.9728292 0.03974152854309923 1 -15 1 -0.8155748 0.07542024 3.7289044851597826 0 -16 0 -1.5951606 0.017058393 0.024822381042273359 0 -17 0 -1.8490299 0.01037536 0.015046673469226494 0 -18 1 2.6272666 0.98698443 0.018900766460410914 1 -19 0 -2.0090678 0.0075727217 0.010966704536312899 0 -20 1 1.6119726 0.9099426 0.13615251047667312 1 -21 1 1.4732794 0.88468856 0.17675843349200604 1 -22 0 -1.4967335 0.020663453 0.030123371045923587 0 +0 0 -1.6781839 0.0069645904 0.010082932679683302 0 +1 0 1.3120825 0.88653606 3.1396942501860932 1 +2 0 -1.2799562 0.017539484 0.025528667293000074 0 +3 0 2.129505 0.9815428 5.7596744891074314 1 +4 0 -1.733461 0.0061226506 0.0088602695578320623 0 +5 1 3.2187092 0.9985419 0.002105140929137923 1 +6 0 0.8574975 0.7289475 1.8833558892994415 1 +7 0 -1.3869908 0.013697766 0.019898293097052152 0 +8 0 -1.5963612 0.008426138 0.012207852857983661 0 +9 0 -1.1096888 0.025928956 0.037901094886052082 0 +10 0 -0.9912352 0.03395389 0.049836043729202285 0 +11 0 -1.2701744 0.017939368 0.026115996614060374 0 +12 1 0.4021077 0.4802237 1.0582215332285976 1 +13 0 -0.90570307 0.04118861 0.060681045880743613 0 +14 1 2.1833656 0.9836982 0.023712348072192725 1 +15 1 -0.569199 0.08642988 3.5323260471396378 0 +16 0 -1.4413924 0.012076221 0.017528357214092197 0 +17 0 -1.6143917 0.0080799395 0.01170423737885466 0 +18 1 2.3743236 0.9895233 0.015194428784074532 1 +19 0 -1.7312822 0.006153836 0.0089055383589370304 0 +20 1 1.5163187 0.9265604 0.11004306714535973 1 +21 1 1.979719 0.9739729 0.038046439238538551 1 +22 0 -1.3621383 0.014508282 0.021084347628400252 0 23 1 ? ? ? 0 -24 0 -1.5333067 0.01924419 0.028034118479806358 0 -25 1 1.9215719 0.94919187 0.075228353822425834 1 -26 0 -0.6941621 0.09404353 0.14248636335705839 0 -27 0 -1.6840487 0.014338439 0.020835730799420298 0 -28 0 -1.3867699 0.025575617 0.03737786115106672 0 -29 0 -0.8064519 0.0766929 0.1151175166592759 0 -30 0 -1.0486952 0.048843503 0.072245362767655683 0 -31 0 -1.6512232 0.015289166 0.022227963838514678 0 -32 1 0.25331846 0.4050943 1.3036703247614843 1 -33 0 -1.5938717 0.01710135 0.024885430797322208 0 -34 0 -1.3701667 0.026410026 0.038613784293999399 0 -35 0 -1.3867699 0.025575617 0.03737786115106672 0 -36 1 1.8522056 0.9421248 0.086009938101436673 1 -37 0 -2.0591261 0.0068612704 0.0099328358258817841 0 -38 1 3.4026167 0.99717873 0.0040759798270225368 1 -39 1 1.302867 0.84544253 0.24222139993538136 1 +24 0 -1.3507648 0.014894797 0.021650290712265013 0 +25 1 1.402212 0.9061323 0.14220641883002672 1 +26 0 -0.7339115 0.060399428 0.089880503283842381 0 +27 0 -1.5088496 0.010326769 0.014975838511155387 0 +28 0 -1.2701744 0.017939368 0.026115996614060374 0 +29 0 -0.9144431 0.040386368 0.059474441974954127 0 +30 0 -1.0090464 0.032609552 0.047829802400531349 0 +31 0 -1.4515353 0.011795582 0.017118589545051147 0 +32 1 1.2101234 0.8601601 0.21732286316814953 1 +33 0 -1.5267994 0.009905118 0.014361307209470462 0 +34 0 -1.292162 0.017052773 0.024814132997407166 0 +35 0 -1.2701744 0.017939368 0.026115996614060374 0 +36 1 2.203938 0.9844544 0.022603722644415333 1 +37 0 -1.5366075 0.009681969 0.01403618773745351 0 +38 1 2.4723532 0.9916579 0.012085567744344158 1 +39 1 1.1451632 0.84080243 0.25016125415492318 1 40 0 ? ? ? 0 -41 1 -0.23856218 0.20411506 2.2925454429216874 0 -42 1 2.8029609 0.99078166 0.013360924200893591 1 -43 1 -1.810855 0.011183086 6.4825378505935216 0 -44 1 1.2626739 0.83472645 0.2606246036558294 1 -45 0 -1.4895467 0.02095416 0.030551684942010555 0 -46 1 3.916806 0.9989816 0.0014699964392690457 1 -47 0 -1.263222 0.03245402 0.047597873731249518 0 -48 0 -1.6886145 0.014210899 0.020649064923297558 0 -49 1 3.300146 0.9965445 0.0049938943284022781 1 -50 1 0.6495075 0.5992249 0.73883046026718302 1 -51 1 1.0815775 0.77902395 0.36026042092624549 1 -52 1 1.4402541 0.87782973 0.18798696238174045 1 -53 1 -0.298494 0.18546546 2.4307776005567865 0 -54 1 0.25865722 0.40765104 1.2945934099554417 1 -55 1 1.6522897 0.91628987 0.1261240314505572 1 -56 1 3.8028905 0.9987235 0.001842766615230386 1 -57 1 -2.8152497 0.001537523 9.3451762929958502 0 -58 1 0.3947263 0.47413352 1.0766346994390334 1 -59 1 0.096370675 0.33273125 1.5875707381331534 1 -60 1 1.6592671 0.9173462 0.12446182761064933 1 -61 0 -0.9575624 0.057968486 0.086152771582063581 0 -62 1 2.0682306 0.9615321 0.056593049357335799 1 -63 1 0.27973998 0.41779524 1.259132036180967 1 -64 0 -1.263222 0.03245402 0.047597873731249518 0 -65 1 -1.9910125 0.0078469105 6.9936595382121061 0 -66 0 -1.8490299 0.01037536 0.015046673469226494 0 -67 1 1.7589653 0.9311664 0.10288907740228966 1 -68 1 1.6836334 0.92094064 0.11881992946846752 1 -69 0 -1.363128 0.026771704 0.03914982815834981 0 -70 0 -1.3740007 0.026215024 0.038324853100870006 0 -71 1 -0.40530378 0.1555408 2.684635085100564 0 -72 0 -1.0371907 0.04991555 0.073872336739382177 0 -73 1 2.6706092 0.9880447 0.017351813247254332 1 -74 1 0.5221059 0.5372601 0.89630735495507363 1 -75 0 -1.1163957 0.042964876 0.06335622061882179 0 -76 0 -0.91849536 0.062351767 0.092881310368226153 0 -77 0 -0.7775572 0.08085469 0.12163513850155845 0 -78 0 -1.0141383 0.052131098 0.077240558097625131 0 -79 0 -1.8685613 0.009984706 0.014477281827753447 0 -80 0 -1.6484528 0.015372191 0.022349607923170395 0 -81 0 -1.5205013 0.019729905 0.028748783134744417 0 -82 0 -1.4972845 0.02064133 0.030090781781786324 0 -83 0 -2.4856832 0.0029535654 0.0042673993513016309 0 -84 1 2.367095 0.97837436 0.031541495919643385 1 -85 1 0.9574349 0.73371196 0.44671429670055623 1 -86 1 1.9397789 0.950907 0.07262385664276523 1 -87 1 3.411163 0.99722606 0.0040075111664453775 1 -88 0 -1.8490299 0.01037536 0.015046673469226494 0 -89 0 -1.8165104 0.011059615 0.016044539131928545 0 -90 0 -1.5333067 0.01924419 0.028034118479806358 0 -91 0 -1.1862181 0.037612904 0.055310796164145914 0 -92 0 -1.8490299 0.01037536 0.015046673469226494 0 -93 0 -1.263222 0.03245402 0.047597873731249518 0 -94 0 -1.6512232 0.015289166 0.022227963838514678 0 -95 0 -1.5333067 0.01924419 0.028034118479806358 0 -96 0 -1.08194 0.045867078 0.067737830183898132 0 -97 0 -1.9328692 0.00879855 0.012749796760162687 0 -98 1 0.6660452 0.6070831 0.72003412523607846 1 -99 1 2.9950843 0.9936863 0.0091375935158915338 1 -100 1 -0.3330408 0.17532746 2.5118760838074556 0 -101 1 2.04923 0.96011233 0.058724883625953142 1 -102 0 -1.7038524 0.01379329 0.020038025642657626 0 -103 1 -3.4719143 0.00041797833 11.224284237787918 0 -104 1 0.3974993 0.47550628 1.0724637124999739 1 -105 1 -0.6976452 0.093456045 3.4195682084042782 0 -106 1 4.220067 0.999442 0.00080527548897498544 1 -107 1 3.704927 0.9984499 0.0022380254150517536 1 -108 0 -0.92559916 0.061532345 0.091621072045217966 0 -109 1 2.5543644 0.98498803 0.021821896941046508 1 -110 0 -0.47045028 0.13929978 0.21641725751686047 0 -111 1 1.7770543 0.9334328 0.099381904687069339 1 -112 1 1.859989 0.9429616 0.084729113706137879 1 -113 1 1.3793039 0.86424893 0.21048118020464698 1 -114 0 -0.030115763 0.27949655 0.4729227575087514 0 -115 0 0.28654557 0.4210852 0.78857707681024869 1 -116 0 -0.39117327 0.15926108 0.25027023007407256 0 -117 1 2.5607088 0.98517317 0.021550762723567849 1 -118 0 -1.2990656 0.030292457 0.044378388658046698 0 -119 0 -0.9705077 0.056580942 0.084029350045610643 0 -120 0 -1.3714294 0.026345644 0.038518383766964329 0 -121 0 -1.041645 0.04949785 0.073238205051276137 0 -122 1 2.1452384 0.96680444 0.048703989117765592 1 -123 1 0.842631 0.68689173 0.54184537031895141 1 -124 1 1.3995808 0.86890274 0.20273339108362537 1 -125 0 -1.263222 0.03245402 0.047597873731249518 0 -126 1 2.03515 0.959028 0.060355149351711469 1 -127 0 -1.7557883 0.012458842 0.018087218256857347 0 -128 1 2.2932503 0.97504514 0.036459077746032992 1 -129 0 -5.2199807 1.3007821E-05 1.8766440325640085E-05 0 -130 0 -1.3740007 0.026215024 0.038324853100870006 0 -131 0 -1.6512232 0.015289166 0.022227963838514678 0 -132 1 3.9878006 0.99911535 0.0012768481185713783 1 -133 0 -1.3060904 0.029885478 0.043773027885930245 0 -134 0 -1.9780878 0.008049211 0.011659545744301534 0 -135 0 -1.0882875 0.045318764 0.06690899061946072 0 -136 0 -1.5951606 0.017058393 0.024822381042273359 0 -137 0 -1.2478162 0.03342823 0.049051234404724475 0 -138 0 -1.4559804 0.022365745 0.032633258967300828 0 +41 1 0.5600745 0.5723573 0.80501205696986078 1 +42 1 2.5724957 0.9933931 0.0095633352592990016 1 +43 1 -1.6388749 0.007632329 7.0336608986855413 0 +44 1 1.950462 0.97217506 0.040711968392278808 1 +45 0 -1.2939029 0.016984442 0.024713845246546313 0 +46 1 2.7204158 0.9953214 0.0067656423594761395 1 +47 0 -1.1645964 0.022866543 0.033372476346272044 0 +48 0 -1.733461 0.0061226506 0.0088602695578320623 0 +49 1 2.6309288 0.9942347 0.008341666299243513 1 +50 1 0.953594 0.7711385 0.3749381169688733 1 +51 1 0.716872 0.65911514 0.60139759452349939 1 +52 1 1.7343138 0.95463014 0.066986213684542981 1 +53 1 0.5582192 0.5712915 0.80770101415814732 1 +54 1 0.9254062 0.75925815 0.39733760328572004 1 +55 1 1.7505497 0.9562517 0.064537716295891306 1 +56 1 2.9442253 0.9972273 0.0040057003265824807 1 +57 1 -1.7144908 0.00639953 7.2878183344942826 0 +58 1 0.5255977 0.55245274 0.85607703360198317 1 +59 1 0.46923366 0.5195761 0.94459310098026461 1 +60 1 1.3479917 0.8947388 0.16046152592582869 1 +61 0 -0.9775615 0.03502207 0.051432146151499443 0 +62 1 2.0915747 0.9798596 0.02935306255578557 1 +63 1 0.4798136 0.52576894 0.92749919002493719 1 +64 0 -1.1645964 0.022866543 0.033372476346272044 0 +65 1 -0.5514082 0.08978313 3.4774117643602511 0 +66 0 -1.6143917 0.0080799395 0.01170423737885466 0 +67 1 1.658551 0.9462784 0.079663410792187223 1 +68 1 2.227339 0.98527265 0.021405090590478744 1 +69 0 -1.2361025 0.019403344 0.028268253118601601 0 +70 0 -1.2991996 0.0167782 0.024411192647472865 0 +71 1 0.81082004 0.7067755 0.5006760920132769 1 +72 0 -0.8947861 0.042212106 0.062221894635322759 0 +73 1 2.3573442 0.9891021 0.015808607404777096 1 +74 1 0.9923081 0.7867729 0.34598081689316063 1 +75 0 -1.0630057 0.028843692 0.04222457756985639 0 +76 0 -1.0024403 0.033102054 0.048564470732650404 0 +77 0 -0.8211149 0.0497804 0.073667127519330286 0 +78 0 -1.0020932 0.03312813 0.048603380435959576 0 +79 0 -1.6449203 0.007525646 0.010898272248269593 0 +80 0 -1.2999079 0.016750805 0.024370994990783851 0 +81 0 -1.502269 0.010485761 0.015207627050754428 0 +82 0 -1.2943019 0.01696882 0.024690918234161204 0 +83 0 -1.9676862 0.0035432624 0.0051209248069075267 0 +84 1 2.2352636 0.98554003 0.021013620520804063 1 +85 1 1.6066236 0.93973744 0.089670368033876791 1 +86 1 1.3669872 0.89886296 0.15382691791336026 1 +87 1 2.6054437 0.99388164 0.0088540373509255182 1 +88 0 -1.6143917 0.0080799395 0.01170423737885466 0 +89 0 -1.7278817 0.0062028244 0.0089766531940858735 0 +90 0 -1.3507648 0.014894797 0.021650290712265013 0 +91 0 -1.2589346 0.018409912 0.026807413333891101 0 +92 0 -1.6143917 0.0080799395 0.01170423737885466 0 +93 0 -1.1645964 0.022866543 0.033372476346272044 0 +94 0 -1.4515353 0.011795582 0.017118589545051147 0 +95 0 -1.3507648 0.014894797 0.021650290712265013 0 +96 0 -1.1702797 0.022570498 0.032935444843664126 0 +97 0 -1.6781839 0.0069645904 0.010082932679683302 0 +98 1 1.5450971 0.9310246 0.10310878994653579 1 +99 1 2.6956 0.9950423 0.007170202978073986 1 +100 1 0.6198605 0.60629034 0.72191925766050979 1 +101 1 1.1729019 0.8493217 0.23561694483900111 1 +102 0 -1.5192053 0.010081387 0.014618177400369553 0 +103 1 -2.3785334 0.0013543374 9.5281970493323644 0 +104 1 1.5962429 0.9383433 0.091812275401115598 1 +105 1 -0.28627953 0.15521893 2.6876235890071394 0 +106 1 3.2182438 0.99854034 0.002107379970098573 1 +107 1 2.5092354 0.99234414 0.0110875665907602 1 +108 0 -0.9480848 0.037435967 0.055045578781261498 0 +109 1 2.3313842 0.9884257 0.016795613091125969 1 +110 0 -0.5008621 0.099957705 0.15193529565530761 0 +111 1 1.6308732 0.9428799 0.084854053279916422 1 +112 1 1.9594742 0.97274137 0.039871825757239431 1 +113 1 1.9867876 0.97439003 0.037428723421636806 1 +114 0 -0.20431276 0.18213779 0.29007028577789618 0 +115 0 -0.1813462 0.19030243 0.30454494987791275 0 +116 0 -0.15897372 0.19852225 0.31926563361034355 0 +117 1 2.2666762 0.9865536 0.01953064568963547 1 +118 0 -1.2926577 0.017033286 0.024785531756646863 0 +119 0 -0.9322239 0.03880024 0.057091805987184267 0 +120 0 -1.3744366 0.014101455 0.020488902786296271 0 +121 0 -0.9642741 0.036091045 0.053031209549174563 0 +122 1 2.390261 0.9899039 0.014639580789807639 1 +123 1 0.9936051 0.78728294 0.34504587192633179 1 +124 1 1.5873636 0.93712693 0.093683619669270399 1 +125 0 -1.1645964 0.022866543 0.033372476346272044 0 +126 1 2.0889175 0.9797362 0.029534734759745728 1 +127 0 -1.539121 0.009625588 0.013954053758422826 0 +128 1 1.9256496 0.97055626 0.043116252425623572 1 +129 0 -4.4531813 1.0431945E-05 1.5050194045784157E-05 0 +130 0 -1.2991996 0.0167782 0.024411192647472865 0 +131 0 -1.4515353 0.011795582 0.017118589545051147 0 +132 1 2.9187195 0.99705684 0.0042523403963686647 1 +133 0 -1.2237993 0.019960253 0.029087833149312455 0 +134 0 -1.7414937 0.006009029 0.0086953482070396826 0 +135 0 -0.9789426 0.034912724 0.051268679045739951 0 +136 0 -1.4413924 0.012076221 0.017528357214092197 0 +137 0 -1.1901796 0.021563126 0.031449319333433555 0 +138 0 -1.3563639 0.014703265 0.021369818969647715 0 139 0 ? ? ? 0 -140 0 -1.2478162 0.03342823 0.049051234404724475 0 -141 0 -1.1315987 0.041740835 0.061512203112918669 0 -142 1 0.8102725 0.6729129 0.57150832571919963 1 -143 0 0.28654557 0.4210852 0.78857707681024869 1 -144 0 -1.3867699 0.025575617 0.03737786115106672 0 +140 0 -1.1901796 0.021563126 0.031449319333433555 0 +141 0 -1.093269 0.02691992 0.039369557847489002 0 +142 1 1.1206626 0.8329565 0.26368695237625217 1 +143 0 -0.1813462 0.19030243 0.30454494987791275 0 +144 0 -1.2701744 0.017939368 0.026115996614060374 0 145 0 ? ? ? 0 -146 1 1.1618232 0.8052274 0.31253183228434933 1 -147 0 -1.4292369 0.023556536 0.034391580380782348 0 -148 0 -3.6192281 0.00031202592 0.00045022849358622251 0 -149 1 1.776138 0.9333197 0.099556766148784281 1 -150 0 -1.0634518 0.047500372 0.070209565626258783 0 -151 1 0.7131937 0.6291722 0.66847315468026391 1 -152 1 2.9240558 0.9927372 0.01051627834880473 1 -153 0 -0.6542493 0.10101492 0.15363092195218497 0 -154 0 -1.0016527 0.053369574 0.079126802248061193 0 -155 1 2.9033995 0.99243546 0.01095481763375002 1 -156 0 -0.7971184 0.07801529 0.11718526969874911 0 -157 0 -1.6512232 0.015289166 0.022227963838514678 0 +146 1 0.8548057 0.72769785 0.45858854809018135 1 +147 0 -1.3318049 0.015561786 0.022627432693713499 0 +148 0 -2.89515 0.00040340194 0.00058210339218496305 0 +149 1 2.325733 0.98827296 0.017018519883359031 1 +150 0 -0.9912352 0.03395389 0.049836043729202285 0 +151 1 1.0926716 0.823618 0.27995274390709196 1 +152 1 2.8430867 0.9964874 0.0050765620383010183 1 +153 0 -0.77696645 0.054914877 0.081483817129461844 0 +154 0 -0.9823508 0.034644295 0.050867463428077943 0 +155 1 2.0343342 0.97703123 0.033523416027709606 1 +156 0 -0.8570405 0.04594131 0.067850074097125401 0 +157 0 -1.4515353 0.011795582 0.017118589545051147 0 158 0 ? ? ? 0 -159 1 3.294172 0.9965034 0.0050533490193179435 1 -160 1 2.8776653 0.9920421 0.011526712997837561 1 -161 0 -1.634767 0.01578888 0.022960277091634333 0 -162 0 -1.7557883 0.012458842 0.018087218256857347 0 -163 0 -1.0945712 0.044782102 0.066098226293311441 0 +159 1 3.1834314 0.9984163 0.002286600634474716 1 +160 1 2.699434 0.9950865 0.007106167352169482 1 +161 0 -1.4432739 0.01202367 0.017451616357808803 0 +162 0 -1.539121 0.009625588 0.013954053758422826 0 +163 0 -0.9426928 0.037894506 0.05573300201932449 0 164 0 ? ? ? 0 -165 0 -1.248234 0.033401437 0.049011245066884986 0 -166 1 3.3694162 0.9969871 0.0043532507609412739 1 -167 1 0.20830406 0.3837536 1.3817478189007735 1 -168 0 -1.7557883 0.012458842 0.018087218256857347 0 -169 0 -0.5321295 0.12525679 0.19306853680046673 0 -170 0 -1.2478162 0.03342823 0.049051234404724475 0 -171 0 -1.5333067 0.01924419 0.028034118479806358 0 -172 0 -1.263222 0.03245402 0.047597873731249518 0 -173 1 3.5792365 0.99801147 0.0028716987760815499 1 -174 1 2.346489 0.9774917 0.032843676743388284 1 -175 1 2.9308467 0.99283373 0.0103759600687816 1 -176 0 -1.6512232 0.015289166 0.022227963838514678 0 -177 1 1.8471723 0.9415775 0.086848257809908569 1 -178 0 -1.8490299 0.01037536 0.015046673469226494 0 -179 1 1.2162164 0.8216072 0.28347931869695053 1 -180 0 -1.0634518 0.047500372 0.070209565626258783 0 -181 0 -1.0016527 0.053369574 0.079126802248061193 0 -182 0 -2.0090678 0.0075727217 0.010966704536312899 0 -183 1 4.06255 0.9992373 0.0011007648308390006 1 -184 1 2.5696895 0.9854313 0.021172778773481976 1 -185 0 -1.1818035 0.037931435 0.055788378507033477 0 -186 1 3.4978397 0.99766356 0.00337471791972376 1 -187 1 1.6904732 0.92192364 0.11728083720565548 1 -188 1 3.3030019 0.9965639 0.0049657642254862539 1 -189 0 -0.88927394 0.065830655 0.098243991981556533 0 -190 1 3.7031476 0.99844444 0.0022459489207712442 1 -191 1 2.949461 0.99309194 0.010000805622003756 1 -192 0 -1.6840487 0.014338439 0.020835730799420298 0 -193 0 -1.5333067 0.01924419 0.028034118479806358 0 -194 0 -1.7557883 0.012458842 0.018087218256857347 0 -195 0 -1.8490299 0.01037536 0.015046673469226494 0 -196 0 1.5795784 0.9045319 3.3888373838747854 1 -197 0 -1.6674584 0.014811429 0.021528203691519851 0 -198 0 -1.0016527 0.053369574 0.079126802248061193 0 -199 0 -1.4967335 0.020663453 0.030123371045923587 0 -200 1 2.9047506 0.9924556 0.010925531323154688 1 -201 1 1.2504892 0.8313623 0.26645075543299601 1 -202 0 -1.5333067 0.01924419 0.028034118479806358 0 -203 0 -1.9328692 0.00879855 0.012749796760162687 0 -204 0 -1.5333067 0.01924419 0.028034118479806358 0 -205 1 4.465726 0.9996573 0.00049453487122621187 1 -206 1 2.1158094 0.9648775 0.051582324766482952 1 -207 0 -1.0634518 0.047500372 0.070209565626258783 0 -208 0 -1.0634518 0.047500372 0.070209565626258783 0 -209 0 -1.4718761 0.02168605 0.031630580006271378 0 -210 1 4.2242675 0.99944663 0.00079856443629892514 1 -211 1 3.7463782 0.9985722 0.0020613942106635767 1 -212 0 -1.5333067 0.01924419 0.028034118479806358 0 -213 1 3.864098 0.9988694 0.0016320061922353926 1 -214 1 4.1788635 0.9993944 0.00087393651373078151 1 -215 1 2.6085155 0.98649746 0.019612755909448824 1 -216 0 -1.263222 0.03245402 0.047597873731249518 0 -217 0 -1.5333067 0.01924419 0.028034118479806358 0 -218 1 2.6015017 0.9863107 0.019885880104498949 1 -219 0 -1.671799 0.014686216 0.021344854574419977 0 -220 0 -1.4890165 0.020975763 0.030583518819028915 0 -221 1 1.254181 0.8323873 0.26467309252117266 1 -222 1 0.39700165 0.4752599 1.0732114150594565 1 -223 1 2.0639515 0.9612166 0.057066488324229336 1 -224 1 2.8239417 0.99115443 0.012818232883687034 1 -225 0 -1.263222 0.03245402 0.047597873731249518 0 -226 1 2.706043 0.98884773 0.016179709814207705 1 -227 1 2.9415061 0.99298275 0.010159446290912691 1 -228 0 -1.0634518 0.047500372 0.070209565626258783 0 -229 1 2.8217773 0.99111664 0.012873238983519546 1 -230 1 2.0452352 0.9598075 0.059182985710066707 1 -231 1 2.8742342 0.9919882 0.011605161547021455 1 -232 0 0.006210236 0.29424706 0.50276486334637094 1 -233 1 1.5423936 0.8979641 0.1552702947542865 1 -234 0 -0.078705244 0.2604905 0.4353594275109402 0 +165 0 -1.1487988 0.02370951 0.034617616343790306 0 +166 1 2.614814 0.99401385 0.0086621472227286639 1 +167 1 0.894213 0.7456284 0.42347125150248932 1 +168 0 -1.539121 0.009625588 0.013954053758422826 0 +169 0 -0.7297581 0.06095483 0.090733541113570265 0 +170 0 -1.1901796 0.021563126 0.031449319333433555 0 +171 0 -1.3507648 0.014894797 0.021650290712265013 0 +172 0 -1.1645964 0.022866543 0.033372476346272044 0 +173 1 3.3297312 0.9988759 0.0016226225592894275 1 +174 1 2.1278155 0.9814709 0.026982626716589176 1 +175 1 2.42077 0.99059474 0.013633127701506339 1 +176 0 -1.4515353 0.011795582 0.017118589545051147 0 +177 1 1.8547715 0.9654135 0.05078107768956798 1 +178 0 -1.6143917 0.0080799395 0.01170423737885466 0 +179 1 0.9386198 0.76487917 0.38669624096896438 1 +180 0 -0.9912352 0.03395389 0.049836043729202285 0 +181 0 -0.9823508 0.034644295 0.050867463428077943 0 +182 0 -1.7312822 0.006153836 0.0089055383589370304 0 +183 1 3.0986078 0.9980682 0.0027896742531955186 1 +184 1 2.2394924 0.9856807 0.020807718132405433 1 +185 0 -1.0878154 0.02725713 0.03986959363330124 0 +186 1 2.5142112 0.9924323 0.010959409919804543 1 +187 1 2.3280811 0.9883367 0.016925507358559522 1 +188 1 2.7640774 0.99577504 0.0061082357396573059 1 +189 0 -0.9172725 0.040129878 0.059088883307958362 0 +190 1 3.2636998 0.9986878 0.0018943421757207767 1 +191 1 2.5799434 0.99350685 0.0093981820543892636 1 +192 0 -1.5088496 0.010326769 0.014975838511155387 0 +193 0 -1.3507648 0.014894797 0.021650290712265013 0 +194 0 -1.539121 0.009625588 0.013954053758422826 0 +195 0 -1.6143917 0.0080799395 0.01170423737885466 0 +196 0 1.8039625 0.9612046 4.6879701577868316 1 +197 0 -1.8342128 0.004839934 0.006999500955684238 0 +198 0 -0.9823508 0.034644295 0.050867463428077943 0 +199 0 -1.3621383 0.014508282 0.021084347628400252 0 +200 1 2.7968245 0.99608624 0.0056574405789838267 1 +201 1 1.8831491 0.96756905 0.047563468136429081 1 +202 0 -1.3507648 0.014894797 0.021650290712265013 0 +203 0 -1.6781839 0.0069645904 0.010082932679683302 0 +204 0 -1.3507648 0.014894797 0.021650290712265013 0 +205 1 3.4962368 0.99923915 0.0010980970675094467 1 +206 1 2.118044 0.98104936 0.02760237131832053 1 +207 0 -0.9912352 0.03395389 0.049836043729202285 0 +208 0 -0.9912352 0.03395389 0.049836043729202285 0 +209 0 -1.3031976 0.016624156 0.024185178742296577 0 +210 1 3.6187198 0.99942905 0.00082394614591436185 1 +211 1 3.0823488 0.9979933 0.0028979786274921419 1 +212 0 -1.3507648 0.014894797 0.021650290712265013 0 +213 1 3.4649906 0.99918133 0.0011815745805306933 1 +214 1 3.5886676 0.9993874 0.00088408967442053408 1 +215 1 2.3846316 0.98977107 0.014833223347463561 1 +216 0 -1.1645964 0.022866543 0.033372476346272044 0 +217 0 -1.3507648 0.014894797 0.021650290712265013 0 +218 1 2.4409509 0.9910258 0.013005470428492955 1 +219 0 -1.4784173 0.011082595 0.016078063975794511 0 +220 0 -1.4339974 0.012284978 0.01783324343708843 0 +221 1 2.0782351 0.9792326 0.030276493094875363 1 +222 1 -0.07737707 0.2307442 2.115633722954938 0 +223 1 1.9717562 0.9734951 0.038754339407876488 1 +224 1 2.7387886 0.9955179 0.0064808243929969037 1 +225 0 -1.1645964 0.022866543 0.033372476346272044 0 +226 1 2.7340784 0.9954683 0.0065526930797491778 1 +227 1 2.6128883 0.9939869 0.0087012499032349793 1 +228 0 -0.9912352 0.03395389 0.049836043729202285 0 +229 1 3.0083337 0.9976136 0.0034469492186543172 1 +230 1 1.9576149 0.97262543 0.040043775980895521 1 +231 1 2.6215558 0.99410725 0.00852659370179807 1 +232 0 0.21231177 0.37181324 0.67073455323125875 1 +233 1 1.6255037 0.9421976 0.085898405807962472 1 +234 0 -0.14271492 0.20466164 0.33035933514005172 0 235 0 ? ? ? 0 -236 1 1.9749312 0.95406413 0.067841849565294918 1 -237 1 1.5597814 0.9010837 0.15026696084403549 1 -238 1 3.772911 0.99864537 0.0019556497479046747 1 -239 1 1.5523349 0.8997582 0.15239071797234102 1 -240 0 -0.34522763 0.17185682 0.27204787501701788 0 -241 0 -1.144221 0.04074999 0.060021219192247205 0 -242 0 -1.6512232 0.015289166 0.022227963838514678 0 -243 0 -0.55725086 0.119893864 0.18425057915990742 0 -244 0 -1.5333067 0.01924419 0.028034118479806358 0 -245 0 -1.4520216 0.022538235 0.032887825323412279 0 -246 1 4.365115 0.9995815 0.0006038714600044195 1 -247 1 2.0720396 0.9618108 0.056174927938110866 1 -248 0 -0.7372442 0.087005325 0.13132164857214251 0 +236 1 2.4899552 0.9919927 0.011598573438484453 1 +237 1 1.8671057 0.9663669 0.04935707383960164 1 +238 1 3.2500732 0.9986452 0.0019559080718386458 1 +239 1 1.7880355 0.9597869 0.059213984965420498 1 +240 0 -0.61675787 0.07801633 0.11718690188313385 0 +241 0 -1.2898105 0.017145494 0.024950227666849324 0 +242 0 -1.4515353 0.011795582 0.017118589545051147 0 +243 0 -0.61680037 0.07800916 0.11717567482359262 0 +244 0 -1.3507648 0.014894797 0.021650290712265013 0 +245 0 -1.4143783 0.012856237 0.018667887049155098 0 +246 1 3.4708958 0.9991926 0.0011653089954989735 1 +247 1 1.8080084 0.96155703 0.056555667445950544 1 +248 0 -0.6863086 0.067057736 0.10014029390971453 0 249 0 ? ? ? 0 -250 0 -0.56830096 0.11759834 0.18049258977657712 0 -251 1 2.3991244 0.9796796 0.029618118551442169 1 -252 0 2.34754 0.9775376 5.476342335033781 1 -253 1 2.8029609 0.99078166 0.013360924200893591 1 -254 1 2.0682306 0.9615321 0.056593049357335799 1 -255 1 3.3221998 0.996692 0.0047803436205580575 1 -256 0 -1.2478162 0.03342823 0.049051234404724475 0 -257 0 -1.4967335 0.020663453 0.030123371045923587 0 -258 0 -1.7557883 0.012458842 0.018087218256857347 0 -259 0 2.6591115 0.98777205 6.3536733389587479 1 -260 1 2.6527734 0.98761916 0.017973267230547148 1 -261 1 2.6916418 0.988528 0.016646244790799371 1 -262 1 3.5533743 0.9979069 0.0030228360023964122 1 -263 1 1.9813198 0.9546168 0.067006391333310822 1 -264 1 0.24258165 0.39996815 1.3220429838726082 1 -265 0 -1.1913133 0.03724847 0.054764583973321714 0 -266 1 3.3243403 0.996706 0.0047600687317822781 1 -267 1 1.8868238 0.94576013 0.080453769533599909 1 -268 1 1.2300134 0.8255864 0.27650892708806529 1 -269 0 -1.5333067 0.01924419 0.028034118479806358 0 -270 1 3.0909743 0.99477506 0.0075577612628225299 1 -271 0 -1.9328692 0.00879855 0.012749796760162687 0 -272 1 1.8868238 0.94576013 0.080453769533599909 1 -273 1 0.04118129 0.30886722 1.694941346826375 1 -274 0 -1.5533432 0.018507635 0.026951050158883109 0 +250 0 -0.71524304 0.06293352 0.093776690855145756 0 +251 1 2.2263708 0.9852396 0.021453442681451551 1 +252 0 1.8816319 0.9674572 4.9415167061586338 1 +253 1 2.5724957 0.9933931 0.0095633352592990016 1 +254 1 2.0915747 0.9798596 0.02935306255578557 1 +255 1 2.4503696 0.99122024 0.012722454396564711 1 +256 0 -1.1901796 0.021563126 0.031449319333433555 0 +257 0 -1.3621383 0.014508282 0.021084347628400252 0 +258 0 -1.539121 0.009625588 0.013954053758422826 0 +259 0 2.054097 0.9780491 5.5095760609896551 1 +260 1 2.3941908 0.98999566 0.0145058965859018 1 +261 1 2.719783 0.9953145 0.0067756642764836197 1 +262 1 3.0454602 0.9978123 0.0031596831352694054 1 +263 1 1.9919584 0.97469103 0.036983122461200404 1 +264 1 0.62551486 0.6094525 0.71441434463916276 1 +265 0 -1.1063769 0.026125932 0.038192866091160071 0 +266 1 2.8273284 0.9963556 0.0052673716740332267 1 +267 1 1.4861065 0.9215888 0.11780494462953149 1 +268 1 2.063501 0.97851783 0.031329955272190319 1 +269 0 -1.3507648 0.014894797 0.021650290712265013 0 +270 1 2.5785468 0.9934857 0.0094289088134946672 1 +271 0 -1.6781839 0.0069645904 0.010082932679683302 0 +272 1 1.4861065 0.9215888 0.11780494462953149 1 +273 1 0.1464739 0.3365036 1.5713061766042866 1 +274 0 -1.3854055 0.013748106 0.019971928641696338 0 275 0 ? ? ? 0 -276 0 -1.4967335 0.020663453 0.030123371045923587 0 -277 0 -1.263222 0.03245402 0.047597873731249518 0 -278 0 -1.5333067 0.01924419 0.028034118479806358 0 -279 1 1.2659237 0.8356146 0.25909035788937806 1 -280 0 -1.7557883 0.012458842 0.018087218256857347 0 -281 0 -1.5938717 0.01710135 0.024885430797322208 0 -282 1 3.371817 0.9970014 0.004332550623543748 1 -283 1 2.146642 0.9668938 0.048570668422270487 1 -284 1 3.710564 0.99846715 0.0022131355551405895 1 -285 1 3.3514583 0.9968781 0.0045110128152735553 1 -286 1 3.0392928 0.99421376 0.0083720245973062206 1 -287 0 -1.9780878 0.008049211 0.011659545744301534 0 -288 1 0.6096163 0.5800659 0.78571126891161869 1 -289 1 3.4734914 0.9975481 0.0035416828596485615 1 -290 0 -1.0016527 0.053369574 0.079126802248061193 0 -291 0 -1.5333067 0.01924419 0.028034118479806358 0 +276 0 -1.3621383 0.014508282 0.021084347628400252 0 +277 0 -1.1645964 0.022866543 0.033372476346272044 0 +278 0 -1.3507648 0.014894797 0.021650290712265013 0 +279 1 1.6948717 0.9504492 0.073318532774508657 1 +280 0 -1.539121 0.009625588 0.013954053758422826 0 +281 0 -1.5267994 0.009905118 0.014361307209470462 0 +282 1 2.1658776 0.983027 0.024697079815852833 1 +283 1 2.1300294 0.9815651 0.026844114447266532 1 +284 1 2.7441695 0.9955739 0.0063997172787036332 1 +285 1 3.1651368 0.9983469 0.0023868567904076462 1 +286 1 3.091509 0.99803585 0.0028364586775264165 1 +287 0 -1.7414937 0.006009029 0.0086953482070396826 0 +288 1 0.8335748 0.7177163 0.47851445581376079 1 +289 1 2.7271237 0.9953941 0.0066602436554640647 1 +290 0 -0.9823508 0.034644295 0.050867463428077943 0 +291 0 -1.3507648 0.014894797 0.021650290712265013 0 292 1 ? ? ? 0 -293 1 2.563754 0.9852612 0.021421847810357383 1 +293 1 2.145035 0.9821915 0.025923753074582907 1 294 0 ? ? ? 0 -295 1 2.6352596 0.98718673 0.018605093455754552 1 -296 0 1.2419703 0.828978 2.5477461893002844 1 +295 1 2.1818213 0.98363996 0.023797756394289226 1 +296 0 1.0705519 0.81595176 2.4418441771034214 1 297 0 ? ? ? 0 -298 0 -2.6113403 0.0023029924 0.0033263474492443651 0 -299 1 1.2160475 0.821558 0.28356566770768643 1 -300 1 1.931308 0.950116 0.07382446385610715 1 -301 0 -1.5333067 0.01924419 0.028034118479806358 0 -302 1 2.7389624 0.9895459 0.015161493389214147 1 -303 0 -1.5333067 0.01924419 0.028034118479806358 0 -304 1 2.4927063 0.98306644 0.024639171827518493 1 -305 1 3.2814639 0.99641436 0.0051822766066175301 1 -306 0 -1.5333067 0.01924419 0.028034118479806358 0 -307 0 -1.5333067 0.01924419 0.028034118479806358 0 -308 1 1.7546115 0.93061036 0.10375084892301178 1 -309 0 -0.6772771 0.09693856 0.1471039464232507 0 -310 0 -1.8685613 0.009984706 0.014477281827753447 0 -311 0 -1.0016527 0.053369574 0.079126802248061193 0 -312 1 -0.6179104 0.1077572 3.2141437820433492 0 -313 0 -1.0016527 0.053369574 0.079126802248061193 0 -314 0 -0.8114933 0.075987205 0.1140152656637158 0 +298 0 -2.2336533 0.0019015739 0.0027460028562535481 0 +299 1 1.5479625 0.9314551 0.10244190550452423 1 +300 1 2.019289 0.9762256 0.034713488724118197 1 +301 0 -1.3507648 0.014894797 0.021650290712265013 0 +302 1 3.0965285 0.9980588 0.0028032872441067559 1 +303 0 -1.3507648 0.014894797 0.021650290712265013 0 +304 1 1.9499217 0.9721407 0.040762917938314075 1 +305 1 2.9005282 0.99692893 0.0044374343796938619 1 +306 0 -1.3507648 0.014894797 0.021650290712265013 0 +307 0 -1.3507648 0.014894797 0.021650290712265013 0 +308 1 1.99119 0.9746465 0.03704902743529806 1 +309 0 -0.64283025 0.07372834 0.11049271769888758 0 +310 0 -1.6449203 0.007525646 0.010898272248269593 0 +311 0 -0.9823508 0.034644295 0.050867463428077943 0 +312 1 -0.4606703 0.10876713 3.2006854693651619 0 +313 0 -0.9823508 0.034644295 0.050867463428077943 0 +314 0 -0.818778 0.05004039 0.074061920586461369 0 315 0 ? ? ? 0 -316 1 2.3204832 0.97632736 0.03456313460355704 1 -317 1 3.181556 0.9956312 0.0063166282435822545 1 -318 0 -3.0949686 0.00088296167 0.0012744071330150258 0 -319 0 -0.02775991 0.28043935 0.47481179561763193 0 -320 1 1.4340866 0.87651056 0.19015662169228964 1 +316 1 1.8602372 0.96583915 0.050145154614264852 1 +317 1 2.735732 0.9954858 0.0065273831458453179 1 +318 0 -2.5778427 0.00084891065 0.0012252393249456049 0 +319 0 0.17768885 0.3530457 0.62826429356532831 1 +320 1 1.847582 0.96484584 0.051629649057253896 1 321 0 ? ? ? 0 -322 0 -1.7557883 0.012458842 0.018087218256857347 0 -323 1 3.2233448 0.99597764 0.0058147409513934203 1 -324 0 -1.5333067 0.01924419 0.028034118479806358 0 -325 0 -0.54311305 0.122887164 0.18916564565503 0 -326 1 0.9645734 0.7364716 0.44129821508474998 1 -327 0 -1.263222 0.03245402 0.047597873731249518 0 -328 1 2.0885391 0.9629959 0.054398458918376795 1 -329 1 1.274679 0.8379882 0.25499816968620692 1 -330 1 1.5122968 0.89235806 0.16430537751874952 1 -331 0 -2.1554403 0.0056739375 0.0082090724608937047 0 -332 0 -0.79770106 0.077932134 0.1170551555690699 0 -333 1 2.7442825 0.98965454 0.015003083776305881 1 -334 1 3.0678036 0.99453044 0.007912566899282639 1 -335 0 -1.0016527 0.053369574 0.079126802248061193 0 -336 1 3.0946832 0.99481314 0.0075025252527491978 1 -337 0 -1.5333067 0.01924419 0.028034118479806358 0 -338 0 -0.8114933 0.075987205 0.1140152656637158 0 -339 1 2.847133 0.9915491 0.012243917489372782 1 -340 1 1.7296926 0.92734706 0.10881871954224319 1 -341 0 -1.5333067 0.01924419 0.028034118479806358 0 -342 0 -1.1315987 0.041740835 0.061512203112918669 0 -343 0 -1.0016527 0.053369574 0.079126802248061193 0 -344 1 1.0678388 0.77429307 0.3690483741243904 1 -345 0 -1.0016527 0.053369574 0.079126802248061193 0 -346 0 -0.7760705 0.081074305 0.12197988583661425 0 -347 0 -0.279373 0.19126847 0.3062672429053202 0 -348 1 0.22655053 0.39235517 1.3497678703351699 1 -349 1 1.6726629 0.9193403 0.12132909164656115 1 -350 0 -1.2832505 0.031228447 0.04577159301446896 0 -351 0 -1.6512232 0.015289166 0.022227963838514678 0 -352 0 0.20322987 0.3813742 0.69286111571576769 1 -353 1 4.320384 0.9995427 0.00065987633659195992 1 -354 0 -1.263222 0.03245402 0.047597873731249518 0 -355 0 -1.8602117 0.010149893 0.014718020974524015 0 -356 1 -0.24053709 0.20347889 2.2970489820838358 0 -357 1 2.108741 0.9643988 0.052298237104353419 1 -358 1 1.7299445 0.92738074 0.10876632899873379 1 -359 1 1.3498648 0.85724515 0.22222026264341668 1 -360 1 3.109418 0.9949619 0.0072867879537769579 1 -361 1 3.1558468 0.9954035 0.0066465942225404674 1 -362 0 -1.0647125 0.04738726 0.070038253249742374 0 -363 0 0.78975993 0.6638877 1.5729846683153197 1 -364 0 -1.6512232 0.015289166 0.022227963838514678 0 -365 0 -1.3867699 0.025575617 0.03737786115106672 0 -366 1 3.859608 0.9988593 0.0016466413379224961 1 -367 1 4.1751122 0.99938995 0.00088038978555527749 1 -368 0 -0.8064519 0.0766929 0.1151175166592759 0 -369 0 -0.6032179 0.1105938 0.16908563845226129 0 -370 0 -1.0435581 0.049319476 0.072967489570264304 0 -371 0 -0.8064519 0.0766929 0.1151175166592759 0 -372 0 -1.4559804 0.022365745 0.032633258967300828 0 -373 0 -1.6019248 0.01683469 0.02449408130257406 0 -374 0 -1.3701667 0.026410026 0.038613784293999399 0 -375 0 -1.0016527 0.053369574 0.079126802248061193 0 -376 0 -1.263222 0.03245402 0.047597873731249518 0 -377 0 -0.8114933 0.075987205 0.1140152656637158 0 -378 0 -1.5247288 0.01956824 0.028510875816427802 0 -379 0 -1.3874888 0.025540072 0.037325235770265495 0 -380 0 -1.0016527 0.053369574 0.079126802248061193 0 -381 1 2.298771 0.97531044 0.036066587890400438 1 -382 0 -0.8713332 0.06805511 0.10168344758551041 0 -383 0 -1.1315987 0.041740835 0.061512203112918669 0 -384 0 -1.1315987 0.041740835 0.061512203112918669 0 -385 0 -0.8080883 0.07646318 0.11475861105265454 0 -386 1 2.6133997 0.986626 0.019424746096498028 1 -387 0 -1.0279458 0.050793152 0.075205586534615565 0 -388 0 -1.2108078 0.035885204 0.052723157670906254 0 -389 0 -1.534474 0.019200502 0.027969854215552061 0 -390 0 -1.2228806 0.035065167 0.051496581294782871 0 -391 1 3.6042433 0.9981076 0.0027327250953220411 1 -392 0 -1.4967335 0.020663453 0.030123371045923587 0 -393 0 -0.56838185 0.11758168 0.18046535235209679 0 -394 0 -0.65670073 0.10057384 0.15292324558231868 0 -395 0 -1.4967335 0.020663453 0.030123371045923587 0 -396 0 -1.7557883 0.012458842 0.018087218256857347 0 -397 0 -1.3522246 0.02734149 0.039994716691546361 0 -398 0 -0.90949297 0.06340482 0.094502482696951939 0 -399 0 -0.43538642 0.147857 0.2308325345964948 0 -400 1 2.7999268 0.9907265 0.013441295273746242 1 -401 0 -1.2478162 0.03342823 0.049051234404724475 0 -402 0 -0.54779065 0.12188976 0.18752602870834698 0 -403 0 -0.221742 0.20959322 0.33933277549366375 0 -404 0 -0.35944894 0.16787587 0.26512934194717935 0 -405 0 -1.263222 0.03245402 0.047597873731249518 0 -406 0 -1.0413047 0.049529646 0.073286465610083271 0 -407 0 -1.263222 0.03245402 0.047597873731249518 0 -408 0 -0.22472577 0.2086136 0.33754582997502025 0 -409 0 -1.3701667 0.026410026 0.038613784293999399 0 -410 0 -1.263222 0.03245402 0.047597873731249518 0 +322 0 -1.539121 0.009625588 0.013954053758422826 0 +323 1 2.199418 0.98429126 0.022842817276452352 1 +324 0 -1.3507648 0.014894797 0.021650290712265013 0 +325 0 -0.6375243 0.074583024 0.11182453037541937 0 +326 1 1.295006 0.88244313 0.18042478952012911 1 +327 0 -1.1645964 0.022866543 0.033372476346272044 0 +328 1 2.0094645 0.9756847 0.03551308621739644 1 +329 1 1.6505216 0.9453126 0.081136580793356547 1 +330 1 1.7133101 0.95244735 0.070288743044948962 1 +331 0 -1.799816 0.005244585 0.0075862478843354486 0 +332 0 -0.8737104 0.04425717 0.065305625103000406 0 +333 1 2.031276 0.9768697 0.033761950631468154 1 +334 1 2.4881847 0.99195963 0.011646684660284796 1 +335 0 -0.9823508 0.034644295 0.050867463428077943 0 +336 1 2.272325 0.98672825 0.019275279652746082 1 +337 0 -1.3507648 0.014894797 0.021650290712265013 0 +338 0 -0.818778 0.05004039 0.074061920586461369 0 +339 1 2.3117535 0.9878867 0.017582466003907945 1 +340 1 2.0182006 0.97616625 0.034801224553984038 1 +341 0 -1.3507648 0.014894797 0.021650290712265013 0 +342 0 -1.093269 0.02691992 0.039369557847489002 0 +343 0 -0.9823508 0.034644295 0.050867463428077943 0 +344 1 1.6436236 0.94446987 0.082423322985509084 1 +345 0 -0.9823508 0.034644295 0.050867463428077943 0 +346 0 -0.88080233 0.043558676 0.06425163010703927 0 +347 0 -0.49027798 0.10221409 0.15555664119534018 0 +348 1 -0.5628029 0.087622166 3.5125603168499318 0 +349 1 1.3887984 0.903421 0.14652966967074496 1 +350 0 -1.450688 0.011818778 0.017152453443940112 0 +351 0 -1.4515353 0.011795582 0.017118589545051147 0 +352 0 0.18701717 0.35806054 0.63949084598160777 1 +353 1 3.0073948 0.99760836 0.003454534576817379 1 +354 0 -1.1645964 0.022866543 0.033372476346272044 0 +355 0 -1.6558841 0.0073359325 0.010622524254068027 0 +356 1 -0.08291201 0.22844721 2.1300672479033484 0 +357 1 2.468861 0.99158984 0.012184599338694074 1 +358 1 1.9528928 0.9723289 0.04048369052231774 1 +359 1 1.5290972 0.92857456 0.10691033549276339 1 +360 1 3.0919123 0.9980377 0.0028337877027055204 1 +361 1 2.3240068 0.98822594 0.017087173759882699 1 +362 0 -1.05336 0.029484423 0.043176725990535 0 +363 0 0.39460513 0.4758316 0.93189770748309009 1 +364 0 -1.4515353 0.011795582 0.017118589545051147 0 +365 0 -1.2701744 0.017939368 0.026115996614060374 0 +366 1 3.3189533 0.9988471 0.0016642037087210332 1 +367 1 3.2917511 0.99877125 0.0017738011759607783 1 +368 0 -0.9144431 0.040386368 0.059474441974954127 0 +369 0 -0.73914325 0.05970657 0.088817055336067749 0 +370 0 -0.97795826 0.03499062 0.051385129082716016 0 +371 0 -0.9144431 0.040386368 0.059474441974954127 0 +372 0 -1.3563639 0.014703265 0.021369818969647715 0 +373 0 -1.5631027 0.009103745 0.013194077387202177 0 +374 0 -1.292162 0.017052773 0.024814132997407166 0 +375 0 -0.9823508 0.034644295 0.050867463428077943 0 +376 0 -1.1645964 0.022866543 0.033372476346272044 0 +377 0 -0.818778 0.05004039 0.074061920586461369 0 +378 0 -1.7652828 0.005684684 0.0082246649829244688 0 +379 0 -1.1643724 0.02287829 0.033389818672750954 0 +380 0 -0.9823508 0.034644295 0.050867463428077943 0 +381 1 2.5076258 0.9923154 0.011129334780984131 1 +382 0 -0.9560025 0.036772296 0.054051207679614521 0 +383 0 -1.093269 0.02691992 0.039369557847489002 0 +384 0 -1.093269 0.02691992 0.039369557847489002 0 +385 0 -0.84588814 0.047101893 0.069606138578020699 0 +386 1 2.273435 0.98676234 0.01922543189744803 1 +387 0 -0.98739815 0.034250423 0.050278954799326987 0 +388 0 -1.1510141 0.0235895 0.034440284371001842 0 +389 0 -1.3951019 0.0134430295 0.019525730501774138 0 +390 0 -1.248862 0.018841866 0.027442420043830181 0 +391 1 3.1779 0.9983956 0.0023164872647895552 1 +392 0 -1.3621383 0.014508282 0.021084347628400252 0 +393 0 -0.6498556 0.07261056 0.10875279122071219 0 +394 0 -0.7809162 0.05443591 0.080752846198566999 0 +395 0 -1.3621383 0.014508282 0.021084347628400252 0 +396 0 -1.539121 0.009625588 0.013954053758422826 0 +397 0 -1.2739269 0.01778492 0.025889121808373532 0 +398 0 -0.9042717 0.041321438 0.060880923505861495 0 +399 0 -0.64939034 0.0726841 0.10886720567174937 0 +400 1 2.5536447 0.9930964 0.0099943114248139259 1 +401 0 -1.1901796 0.021563126 0.031449319333433555 0 +402 0 -0.5660218 0.08702029 0.13134530122774041 0 +403 0 -0.48567685 0.10320898 0.15715626541102551 0 +404 0 -0.75046104 0.05823311 0.086558090154963088 0 +405 0 -1.1645964 0.022866543 0.033372476346272044 0 +406 0 -1.0087458 0.032631814 0.047863003193122228 0 +407 0 -1.1645964 0.022866543 0.033372476346272044 0 +408 0 -0.47484383 0.105585285 0.16098417039215199 0 +409 0 -1.292162 0.017052773 0.024814132997407166 0 +410 0 -1.1645964 0.022866543 0.033372476346272044 0 411 0 ? ? ? 0 -412 1 1.5375534 0.89708036 0.15669086510576993 1 -413 0 -1.7772654 0.011945016 0.017336766340184002 0 -414 1 2.7025042 0.98877 0.016293111596682109 1 -415 0 0.87615424 0.70102394 1.7418981092022934 1 -416 1 1.5371778 0.89701146 0.15680167991591476 1 -417 0 -1.263222 0.03245402 0.047597873731249518 0 -418 0 -0.7854102 0.07970364 0.11982956703589996 0 -419 0 -1.7714506 0.012082028 0.01753683707960332 0 -420 0 -0.22567253 0.20830348 0.33698058747522869 0 -421 1 2.284518 0.97461987 0.037088465988161581 1 -422 0 -0.24295312 0.20270263 0.32681017853524641 0 -423 0 -1.3740007 0.026215024 0.038324853100870006 0 -424 0 -1.2478162 0.03342823 0.049051234404724475 0 -425 1 3.353119 0.99688834 0.0044961760643144281 1 -426 0 0.10532635 0.33669022 0.59224529144325821 1 -427 1 1.8121941 0.9376385 0.092896253375268759 1 -428 0 -1.263222 0.03245402 0.047597873731249518 0 -429 0 -1.3867699 0.025575617 0.03737786115106672 0 -430 0 -0.36221886 0.16710912 0.26380059418147622 0 -431 0 -2.5259378 0.0027273379 0.0039400922614880526 0 -432 0 -1.4725443 0.021657925 0.031589106728466369 0 -433 0 -0.5699 0.11726934 0.17995478236362938 0 -434 0 2.9324195 0.9928559 7.1290333197455702 1 -435 1 3.0144103 0.99392253 0.0087946853785197069 1 -436 1 1.7865667 0.9345966 0.097584308961605887 1 -437 0 -1.3522246 0.02734149 0.039994716691546361 0 -438 0 -0.8901633 0.06572216 0.098076445677239582 0 -439 0 -1.4468354 0.022766171 0.033224288535319356 0 -440 1 1.2481554 0.8307117 0.26758018209918738 1 -441 0 0.41069767 0.4820448 0.94910073759986457 1 -442 0 -0.23697054 0.20462884 0.33029984379212168 0 -443 0 -0.71370757 0.090789296 0.13731342557756876 0 -444 0 -1.7059597 0.013736499 0.019954949846956591 0 -445 0 -1.1315987 0.041740835 0.061512203112918669 0 -446 0 -1.0016527 0.053369574 0.079126802248061193 0 -447 0 -1.4468354 0.022766171 0.033224288535319356 0 -448 0 -0.56838185 0.11758168 0.18046535235209679 0 -449 1 3.9154434 0.99897885 0.0014739560781831013 1 -450 0 -0.9983365 0.05370316 0.079635285817558701 0 -451 0 -1.4468354 0.022766171 0.033224288535319356 0 -452 0 -1.1156343 0.04302707 0.063449978399446105 0 -453 1 2.5579932 0.9850942 0.021666420637650348 1 -454 0 -0.8438194 0.07160216 0.1071849260372788 0 -455 1 0.19439556 0.3772452 1.4064255974221649 1 -456 1 2.4136403 0.98024535 0.028785199220143737 1 -457 1 1.7859749 0.9345248 0.097695184101882285 1 -458 0 -1.3039718 0.030007659 0.043954738269590315 0 -459 0 -1.1502336 0.04028595 0.059323478815427419 0 -460 0 -1.2832505 0.031228447 0.04577159301446896 0 -461 0 0.12462831 0.3453007 0.61109566782476865 1 -462 0 -1.3592134 0.026974931 0.039451120366340367 0 -463 0 -1.1623768 0.03936417 0.057938475786410405 0 -464 0 -1.3522246 0.02734149 0.039994716691546361 0 -465 1 2.0827668 0.9625854 0.055013569040676613 1 -466 1 2.3157465 0.9761091 0.034885706163314996 1 -467 1 2.9665604 0.99332094 0.009668167149276409 1 -468 0 -1.3522246 0.02734149 0.039994716691546361 0 -469 0 -1.1693764 0.03884206 0.057154576473333961 0 -470 0 -1.0486952 0.048843503 0.072245362767655683 0 -471 0 -1.3592134 0.026974931 0.039451120366340367 0 -472 0 -1.1201133 0.042662434 0.062900373395280623 0 -473 0 -1.3522246 0.02734149 0.039994716691546361 0 -474 0 -1.4468354 0.022766171 0.033224288535319356 0 -475 0 -1.2478162 0.03342823 0.049051234404724475 0 -476 0 -1.2198313 0.035270564 0.051803708408346028 0 -477 0 -1.3522246 0.02734149 0.039994716691546361 0 -478 0 -1.0614573 0.047679838 0.070481417597247603 0 -479 1 4.0574555 0.99922955 0.0011119522791671055 1 -480 0 -1.0274302 0.050842524 0.075280627651705026 0 -481 0 -0.24226339 0.202924 0.3272108021152394 0 -482 1 1.3848407 0.86553335 0.2083386823537168 1 -483 1 3.3480964 0.9968572 0.0045412043494251278 1 -484 0 -1.3039718 0.030007659 0.043954738269590315 0 -485 0 0.1567006 0.35983276 0.64347925304994658 1 -486 0 -1.0486952 0.048843503 0.072245362767655683 0 -487 1 2.857151 0.9917141 0.012003798092893049 1 -488 1 2.1514478 0.96719784 0.048117078997260969 1 -489 1 -0.4095776 0.15442964 2.6949783756609618 0 -490 0 -1.0016527 0.053369574 0.079126802248061193 0 -491 1 3.483061 0.9975942 0.0034750497225520651 1 -492 0 -1.1507912 0.040243175 0.059259180340592646 0 -493 1 4.3156133 0.99953836 0.00066615658888828178 1 -494 0 0.10313905 0.33572114 0.59013908148828809 1 -495 0 -1.0486952 0.048843503 0.072245362767655683 0 -496 0 -0.56838185 0.11758168 0.18046535235209679 0 -497 0 -0.9778609 0.055806745 0.082845917799453384 0 -498 0 -1.5951606 0.017058393 0.024822381042273359 0 -499 0 -1.5951606 0.017058393 0.024822381042273359 0 -500 0 -2.0090678 0.0075727217 0.010966704536312899 0 -501 0 -1.5951606 0.017058393 0.024822381042273359 0 -502 0 -1.5205013 0.019729905 0.028748783134744417 0 -503 0 -1.8490299 0.01037536 0.015046673469226494 0 -504 0 -1.0016527 0.053369574 0.079126802248061193 0 -505 0 -0.5240843 0.12701723 0.19597491478494836 0 -506 1 2.7142437 0.98902583 0.015919893326652064 1 -507 0 -0.05555363 0.26944113 0.45292755732512568 0 -508 0 -1.4468354 0.022766171 0.033224288535319356 0 -509 0 -1.1315987 0.041740835 0.061512203112918669 0 -510 0 -1.0016527 0.053369574 0.079126802248061193 0 -511 0 -1.6840487 0.014338439 0.020835730799420298 0 -512 0 -1.4468354 0.022766171 0.033224288535319356 0 -513 0 -1.0486952 0.048843503 0.072245362767655683 0 -514 1 4.0258093 0.9991796 0.0011840703743562597 1 -515 1 4.467802 0.9996587 0.00049247037333393663 1 -516 0 -0.56838185 0.11758168 0.18046535235209679 0 -517 0 -0.8114933 0.075987205 0.1140152656637158 0 -518 0 -1.0535297 0.048399545 0.071572133693817208 0 -519 1 2.7834463 0.990421 0.013886194559061286 1 -520 0 -1.4650936 0.021973561 0.032054629522002269 0 -521 0 -1.8115587 0.011167648 0.016202150038975436 0 -522 1 0.552055 0.5520044 0.85724833624835539 1 -523 1 2.9438725 0.9930154 0.010111990812413693 1 -524 0 -1.4967335 0.020663453 0.030123371045923587 0 -525 0 -1.1862181 0.037612904 0.055310796164145914 0 -526 0 -1.3522246 0.02734149 0.039994716691546361 0 -527 0 -1.8490299 0.01037536 0.015046673469226494 0 -528 0 -1.1253302 0.042241435 0.062266073262671563 0 -529 0 -1.1507912 0.040243175 0.059259180340592646 0 -530 1 2.5846334 0.98585117 0.020558231735212602 1 -531 0 -1.0413047 0.049529646 0.073286465610083271 0 -532 0 -1.0634518 0.047500372 0.070209565626258783 0 -533 0 -1.4967335 0.020663453 0.030123371045923587 0 -534 0 -1.3867699 0.025575617 0.03737786115106672 0 -535 0 -1.2347207 0.03427849 0.050320882597922667 0 -536 0 -1.9328692 0.00879855 0.012749796760162687 0 -537 0 -1.7772654 0.011945016 0.017336766340184002 0 -538 0 -1.5951606 0.017058393 0.024822381042273359 0 -539 0 -1.765243 0.012230008 0.01775295378736123 0 -540 0 -1.3124877 0.029519463 0.043228814863948577 0 -541 0 -1.2478162 0.03342823 0.049051234404724475 0 -542 0 -0.6871226 0.09524095 0.14439446235674788 0 -543 0 -1.5951606 0.017058393 0.024822381042273359 0 -544 0 -1.0632471 0.04751876 0.070237417164655452 0 -545 0 -1.6840487 0.014338439 0.020835730799420298 0 -546 1 5.1607866 0.99991375 0.00012443481409838889 1 -547 0 -0.69951284 0.0931424 0.14105206297197442 0 -548 0 -0.81884027 0.07496943 0.11242705599045592 0 -549 1 2.09787 0.96365035 0.05341832590425713 1 -550 0 -1.4967335 0.020663453 0.030123371045923587 0 -551 0 -1.5333067 0.01924419 0.028034118479806358 0 -552 0 -1.0966015 0.044610005 0.065838325943911438 0 -553 0 0.89045954 0.7069419 1.770741394512509 1 -554 0 -1.2478162 0.03342823 0.049051234404724475 0 -555 0 0.22780302 0.39294815 0.72010835002504481 1 -556 0 -0.6235364 0.10668806 0.16276404955240184 0 -557 0 -1.2832505 0.031228447 0.04577159301446896 0 -558 0 -1.3867699 0.025575617 0.03737786115106672 0 -559 0 -1.6840487 0.014338439 0.020835730799420298 0 -560 0 -1.9328692 0.00879855 0.012749796760162687 0 -561 0 -1.9328692 0.00879855 0.012749796760162687 0 -562 0 -1.5333067 0.01924419 0.028034118479806358 0 -563 0 -1.4967335 0.020663453 0.030123371045923587 0 -564 0 -1.634767 0.01578888 0.022960277091634333 0 -565 1 4.19453 0.99941295 0.00084717725458209527 1 -566 0 -1.6231902 0.016150022 0.02348975126265276 0 -567 0 -0.86379343 0.06901059 0.10316334230339028 0 -568 1 1.7233721 0.9264971 0.11014163158543923 1 -569 1 1.813396 0.9377779 0.092681850089216797 1 -570 1 1.9594412 0.95269746 0.069909958606557882 1 -571 1 3.8116453 0.9987455 0.001810995610064468 1 -572 0 -1.4967335 0.020663453 0.030123371045923587 0 -573 0 -1.263222 0.03245402 0.047597873731249518 0 -574 1 1.4165332 0.87268895 0.19646056651204846 1 -575 0 -1.7772654 0.011945016 0.017336766340184002 0 -576 0 -1.6840487 0.014338439 0.020835730799420298 0 -577 0 -1.263222 0.03245402 0.047597873731249518 0 -578 0 -1.263222 0.03245402 0.047597873731249518 0 -579 0 -1.5333067 0.01924419 0.028034118479806358 0 -580 0 -1.5345119 0.019199086 0.027967771942356327 0 -581 1 3.6061473 0.9981147 0.0027224727625651994 1 -582 1 4.295886 0.99951994 0.00069274042536410371 1 -583 0 -1.2478162 0.03342823 0.049051234404724475 0 -584 0 -1.4564363 0.022345966 0.032604070755968044 0 -585 0 -1.0016527 0.053369574 0.079126802248061193 0 -586 1 4.2614655 0.99948597 0.00074177985557674631 1 -587 0 -1.4725443 0.021657925 0.031589106728466369 0 -588 1 1.2662572 0.8357055 0.25893343191422191 1 -589 0 -1.4468354 0.022766171 0.033224288535319356 0 -590 1 1.2167286 0.8217561 0.28321779122674878 1 -591 1 3.4773362 0.9975667 0.0035147878724620519 1 -592 1 1.6255758 0.91213125 0.13268666088935557 1 -593 0 -1.3039718 0.030007659 0.043954738269590315 0 -594 1 2.9495676 0.9930934 0.0099987274757224361 1 -595 0 -1.6840487 0.014338439 0.020835730799420298 0 -596 0 -1.4559804 0.022365745 0.032633258967300828 0 -597 0 -1.6103677 0.01655951 0.024090339531806119 0 -598 0 -1.4967335 0.020663453 0.030123371045923587 0 -599 0 -0.22703165 0.20785888 0.33617061706721391 0 -600 0 -1.4967335 0.020663453 0.030123371045923587 0 -601 0 -0.8114933 0.075987205 0.1140152656637158 0 -602 0 -1.5951606 0.017058393 0.024822381042273359 0 -603 1 1.048409 0.76748085 0.3817973413266334 1 -604 1 0.961967 0.7354661 0.44326920594740676 1 -605 1 2.121477 0.96525675 0.051015356106659845 1 -606 0 -1.2808434 0.03137334 0.045987384326269157 0 -607 0 -1.0016527 0.053369574 0.079126802248061193 0 -608 1 2.7387946 0.9895424 0.015166533585610468 1 -609 0 -1.4468354 0.022766171 0.033224288535319356 0 -610 1 1.7062161 0.9241438 0.11381075113590429 1 -611 1 2.5215163 0.9839927 0.023280488428634154 1 -612 1 3.5015972 0.99768084 0.003349722250434129 1 -613 0 -0.6419177 0.10325986 0.15723811996359513 0 -614 0 -0.93617445 0.060331114 0.089775614857687855 0 -615 0 -1.3033465 0.030043809 0.044008506437147613 0 -616 0 -1.4967335 0.020663453 0.030123371045923587 0 +412 1 2.0363054 0.9771348 0.033370457763616988 1 +413 0 -1.5693717 0.008972017 0.013002300614812229 0 +414 1 2.4392855 0.990991 0.013056145008053159 1 +415 0 0.30689213 0.42493635 0.79820645803751411 1 +416 1 1.8404601 0.9642747 0.05248389210428827 1 +417 0 -1.1645964 0.022866543 0.033372476346272044 0 +418 0 -0.75181305 0.058059387 0.086291990357080156 0 +419 0 -1.5696757 0.008965678 0.012993071859889875 0 +420 0 -0.55675495 0.08876322 0.13410211888439968 0 +421 1 2.5090294 0.99234045 0.01109293919476282 1 +422 0 -0.29067564 0.15387128 0.24105094531791205 0 +423 0 -1.2991996 0.0167782 0.024411192647472865 0 +424 0 -1.1901796 0.021563126 0.031449319333433555 0 +425 1 3.239214 0.9986103 0.0020062821229086536 1 +426 0 -0.11818216 0.21419013 0.34774779798335564 0 +427 1 1.8897946 0.9680548 0.046839419055461608 1 +428 0 -1.1645964 0.022866543 0.033372476346272044 0 +429 0 -1.2701744 0.017939368 0.026115996614060374 0 +430 0 -0.52020735 0.09594783 0.14552206944584331 0 +431 0 -2.414472 0.0012449587 0.0017972147370867028 0 +432 0 -1.3365781 0.015391149 0.022377385809384916 0 +433 0 -0.90458447 0.04129238 0.060837196507563335 0 +434 0 2.283834 0.9870773 6.2739481881414472 1 +435 1 2.665773 0.9946849 0.0076885554279966083 1 +436 1 1.8883485 0.9679496 0.046996121899928571 1 +437 0 -1.2739269 0.01778492 0.025889121808373532 0 +438 0 -1.1902663 0.021558836 0.031442994270673805 0 +439 0 -1.345466 0.015078313 0.021919076775756236 0 +440 1 1.640481 0.9440819 0.083016070882084264 1 +441 0 -0.08429576 0.22787549 0.37309457741745217 0 +442 0 -0.71277225 0.06327625 0.094304456309963661 0 +443 0 -0.81749237 0.05018397 0.074279990920843364 0 +444 0 -2.178376 0.002164329 0.0031258505294281202 0 +445 0 -1.093269 0.02691992 0.039369557847489002 0 +446 0 -0.9823508 0.034644295 0.050867463428077943 0 +447 0 -1.345466 0.015078313 0.021919076775756236 0 +448 0 -0.6498556 0.07261056 0.10875279122071219 0 +449 1 3.1127481 0.99813116 0.0026986945233386144 1 +450 0 -1.1903577 0.021554312 0.03143632318589562 0 +451 0 -1.345466 0.015078313 0.021919076775756236 0 +452 0 -1.3235877 0.015859911 0.02306440228474211 0 +453 1 2.2863357 0.9871519 0.018655965107423074 1 +454 0 -0.8849747 0.04315266 0.063639326480036132 0 +455 1 0.4045404 0.4816485 1.0539474079069142 1 +456 1 2.498416 0.9921489 0.011371475458749423 1 +457 1 2.1684096 0.9831258 0.024552051801129977 1 +458 0 -1.2567189 0.01850409 0.02694583994270246 0 +459 0 -1.1557095 0.023337085 0.034067377608879762 0 +460 0 -1.450688 0.011818778 0.017152453443940112 0 +461 0 -0.23229268 0.17256209 0.2732770415347841 0 +462 0 -1.4985532 0.010576606 0.015340083502900718 0 +463 0 -1.1331927 0.024572 0.035892708719434535 0 +464 0 -1.2739269 0.01778492 0.025889121808373532 0 +465 1 2.3888295 0.9898703 0.014688575373785382 1 +466 1 2.5607123 0.9932092 0.0098304941468018315 1 +467 1 2.538726 0.9928523 0.010349023985076632 1 +468 0 -1.2739269 0.01778492 0.025889121808373532 0 +469 0 -1.1284822 0.024838284 0.0362866062628436 0 +470 0 -1.0090464 0.032609552 0.047829802400531349 0 +471 0 -1.4985532 0.010576606 0.015340083502900718 0 +472 0 -1.0586367 0.02913222 0.042653264141859347 0 +473 0 -1.2739269 0.01778492 0.025889121808373532 0 +474 0 -1.345466 0.015078313 0.021919076775756236 0 +475 0 -1.1901796 0.021563126 0.031449319333433555 0 +476 0 -1.1993723 0.021112747 0.030785394008814505 0 +477 0 -1.2739269 0.01778492 0.025889121808373532 0 +478 0 -1.0173602 0.0319998 0.046920749837993742 0 +479 1 2.9515624 0.9972745 0.0039374074538870421 1 +480 0 -1.0007652 0.033228073 0.048752514371979576 0 +481 0 -0.61380696 0.07851579 0.11796865060611543 0 +482 1 2.3910148 0.98992157 0.014613867985955851 1 +483 1 3.05347 0.99785286 0.0031009958418367167 1 +484 0 -1.2567189 0.01850409 0.02694583994270246 0 +485 0 -0.40593615 0.121855065 0.18746902366589546 0 +486 0 -1.0090464 0.032609552 0.047829802400531349 0 +487 1 2.859111 0.99661654 0.0048895740940511591 1 +488 1 1.2700629 0.87623507 0.19061014090578732 1 +489 1 -0.24454847 0.16849497 2.5692225742946695 0 +490 0 -0.9823508 0.034644295 0.050867463428077943 0 +491 1 2.6771986 0.9948247 0.0074857560531488791 1 +492 0 -1.0850155 0.027431844 0.040128739222532787 0 +493 1 3.191591 0.9984463 0.0022432790389827926 1 +494 0 -0.8224173 0.04963605 0.073447984832625129 0 +495 0 -1.0090464 0.032609552 0.047829802400531349 0 +496 0 -0.6498556 0.07261056 0.10875279122071219 0 +497 0 -0.9703342 0.035599668 0.052295946661420956 0 +498 0 -1.4413924 0.012076221 0.017528357214092197 0 +499 0 -1.4413924 0.012076221 0.017528357214092197 0 +500 0 -1.7312822 0.006153836 0.0089055383589370304 0 +501 0 -1.4413924 0.012076221 0.017528357214092197 0 +502 0 -1.502269 0.010485761 0.015207627050754428 0 +503 0 -1.6143917 0.0080799395 0.01170423737885466 0 +504 0 -0.9823508 0.034644295 0.050867463428077943 0 +505 0 -0.62782 0.0761698 0.11430038044627198 0 +506 1 2.561932 0.9932285 0.0098024427362345748 1 +507 0 -0.4079538 0.121349424 0.18663855031007523 0 +508 0 -1.345466 0.015078313 0.021919076775756236 0 +509 0 -1.093269 0.02691992 0.039369557847489002 0 +510 0 -0.9823508 0.034644295 0.050867463428077943 0 +511 0 -1.5088496 0.010326769 0.014975838511155387 0 +512 0 -1.345466 0.015078313 0.021919076775756236 0 +513 0 -1.0090464 0.032609552 0.047829802400531349 0 +514 1 3.1964872 0.998464 0.0022177000993719802 1 +515 1 3.3663228 0.9989683 0.0014891921814529211 1 +516 0 -0.6498556 0.07261056 0.10875279122071219 0 +517 0 -0.818778 0.05004039 0.074061920586461369 0 +518 0 -1.1884776 0.021647535 0.031573785621762419 0 +519 1 2.3274386 0.9883193 0.016950913365000386 1 +520 0 -1.3386577 0.015317383 0.022269305329613825 0 +521 0 -1.6423699 0.007570472 0.01096343427448079 0 +522 1 1.061431 0.8127164 0.29917604441239076 1 +523 1 2.3249602 0.988252 0.017049148332109122 1 +524 0 -1.3621383 0.014508282 0.021084347628400252 0 +525 0 -1.2589346 0.018409912 0.026807413333891101 0 +526 0 -1.2739269 0.01778492 0.025889121808373532 0 +527 0 -1.6143917 0.0080799395 0.01170423737885466 0 +528 0 -1.222067 0.020039914 0.029205106023067182 0 +529 0 -1.0850155 0.027431844 0.040128739222532787 0 +530 1 2.3383179 0.9886103 0.016526117617036748 1 +531 0 -1.0087458 0.032631814 0.047863003193122228 0 +532 0 -0.9912352 0.03395389 0.049836043729202285 0 +533 0 -1.3621383 0.014508282 0.021084347628400252 0 +534 0 -1.2701744 0.017939368 0.026115996614060374 0 +535 0 -1.2976872 0.016836837 0.024497232734276137 0 +536 0 -1.6781839 0.0069645904 0.010082932679683302 0 +537 0 -1.5693717 0.008972017 0.013002300614812229 0 +538 0 -1.4413924 0.012076221 0.017528357214092197 0 +539 0 -1.5653775 0.009055726 0.013124165847331715 0 +540 0 -1.2269062 0.019818159 0.028878675218473749 0 +541 0 -1.1901796 0.021563126 0.031449319333433555 0 +542 0 -0.693379 0.06602737 0.098547827156875059 0 +543 0 -1.4413924 0.012076221 0.017528357214092197 0 +544 0 -1.1992266 0.021119813 0.030795806542074897 0 +545 0 -1.5088496 0.010326769 0.014975838511155387 0 +546 1 3.7707925 0.99960035 0.00057668708094368046 1 +547 0 -0.75701725 0.057395227 0.08527510863237231 0 +548 0 -0.85085833 0.04658126 0.068818113492030639 0 +549 1 2.0672364 0.9787013 0.031059488570502131 1 +550 0 -1.3621383 0.014508282 0.021084347628400252 0 +551 0 -1.3507648 0.014894797 0.021650290712265013 0 +552 0 -1.0869623 0.027310247 0.039948374780675477 0 +553 0 0.5827884 0.58534795 1.2700268715847562 1 +554 0 -1.1901796 0.021563126 0.031449319333433555 0 +555 0 0.24595208 0.39042673 0.71412844186803504 1 +556 0 -0.74523085 0.058909737 0.087594991112892942 0 +557 0 -1.450688 0.011818778 0.017152453443940112 0 +558 0 -1.2701744 0.017939368 0.026115996614060374 0 +559 0 -1.5088496 0.010326769 0.014975838511155387 0 +560 0 -1.6781839 0.0069645904 0.010082932679683302 0 +561 0 -1.6781839 0.0069645904 0.010082932679683302 0 +562 0 -1.3507648 0.014894797 0.021650290712265013 0 +563 0 -1.3621383 0.014508282 0.021084347628400252 0 +564 0 -1.4432739 0.01202367 0.017451616357808803 0 +565 1 3.347941 0.9989229 0.0015547867347443908 1 +566 0 -1.4586738 0.0116019435 0.016835921041087821 0 +567 0 -0.90413857 0.041333813 0.060899547124751631 0 +568 1 1.6824867 0.9490627 0.075424684920597826 1 +569 1 2.201195 0.98435557 0.022748554924236219 1 +570 1 2.1591165 0.98276025 0.025088588293899899 1 +571 1 3.2443902 0.99862707 0.0019820851369754798 1 +572 0 -1.3621383 0.014508282 0.021084347628400252 0 +573 0 -1.1645964 0.022866543 0.033372476346272044 0 +574 1 1.8664072 0.9663136 0.04943662785325996 1 +575 0 -1.5693717 0.008972017 0.013002300614812229 0 +576 0 -1.5088496 0.010326769 0.014975838511155387 0 +577 0 -1.1645964 0.022866543 0.033372476346272044 0 +578 0 -1.1645964 0.022866543 0.033372476346272044 0 +579 0 -1.3507648 0.014894797 0.021650290712265013 0 +580 0 -1.4099677 0.01298823 0.018860806445465748 0 +581 1 2.8160095 0.99625784 0.0054089202266609316 1 +582 1 3.2695198 0.99870557 0.0018686833191647368 1 +583 0 -1.1901796 0.021563126 0.031449319333433555 0 +584 0 -1.9951745 0.0033226945 0.0048016163833107346 0 +585 0 -0.9823508 0.034644295 0.050867463428077943 0 +586 1 3.6056406 0.9994113 0.00084958642800051745 1 +587 0 -1.3365781 0.015391149 0.022377385809384916 0 +588 1 1.3023883 0.884228 0.17750968994628905 1 +589 0 -1.345466 0.015078313 0.021919076775756236 0 +590 1 1.3021371 0.8841677 0.17760811047878092 1 +591 1 2.5060294 0.9922868 0.011170930860985974 1 +592 1 1.6167625 0.94107056 0.087625202008443065 1 +593 0 -1.2567189 0.01850409 0.02694583994270246 0 +594 1 2.3176296 0.9880506 0.017343197122927183 1 +595 0 -1.5088496 0.010326769 0.014975838511155387 0 +596 0 -1.3563639 0.014703265 0.021369818969647715 0 +597 0 -1.4489241 0.01186721 0.017223164197354753 0 +598 0 -1.3621383 0.014508282 0.021084347628400252 0 +599 0 -0.40352866 0.12246078 0.18846449358617576 0 +600 0 -1.3621383 0.014508282 0.021084347628400252 0 +601 0 -0.818778 0.05004039 0.074061920586461369 0 +602 0 -1.4413924 0.012076221 0.017528357214092197 0 +603 1 1.0872475 0.82176167 0.28320805942670013 1 +604 1 0.92737204 0.7601002 0.39573850773437058 1 +605 1 2.410922 0.99037707 0.013950184531539518 1 +606 0 -1.335089 0.015444186 0.022455100139332329 0 +607 0 -0.9823508 0.034644295 0.050867463428077943 0 +608 1 2.4923072 0.9920364 0.011535034409647822 1 +609 0 -1.345466 0.015078313 0.021919076775756236 0 +610 1 1.8076319 0.96152437 0.056604675508105072 1 +611 1 2.143478 0.9821275 0.026017785341218749 1 +612 1 3.4782767 0.9992064 0.0011453430254316165 1 +613 0 -0.7036707 0.06455383 0.096273453979781418 0 +614 0 -0.920561 0.039833732 0.058643841843197694 0 +615 0 -1.2578133 0.018457511 0.026877375124410151 0 +616 0 -1.3621383 0.014508282 0.021084347628400252 0 617 0 ? ? ? 0 -618 0 -1.5951606 0.017058393 0.024822381042273359 0 -619 0 -1.6840487 0.014338439 0.020835730799420298 0 -620 0 -1.4967335 0.020663453 0.030123371045923587 0 -621 0 -1.1992536 0.036687292 0.053923897213354874 0 -622 0 -1.4792686 0.021376872 0.031174716848926146 0 -623 0 -1.0016527 0.053369574 0.079126802248061193 0 -624 0 -1.1642482 0.03922392 0.057727862343103925 0 -625 0 -0.2870816 0.18891248 0.30207050064878366 0 -626 1 1.552842 0.899849 0.15224516977995409 1 -627 0 0.20483771 0.38212755 0.69461905513595268 1 -628 0 -1.1315987 0.041740835 0.061512203112918669 0 -629 0 -1.3522246 0.02734149 0.039994716691546361 0 -630 0 -0.8208248 0.074696675 0.11200171972149169 0 -631 0 -1.6840487 0.014338439 0.020835730799420298 0 -632 0 -1.0016527 0.053369574 0.079126802248061193 0 -633 1 1.5048606 0.8909318 0.16661311999290043 1 -634 0 -1.2478162 0.03342823 0.049051234404724475 0 -635 0 -0.8436483 0.07162473 0.10722000756055611 0 -636 1 1.0291402 0.7605848 0.39481904115324706 1 -637 0 -0.13322082 0.2401886 0.39628673432889994 0 -638 0 -1.3522246 0.02734149 0.039994716691546361 0 -639 0 -1.2832505 0.031228447 0.04577159301446896 0 -640 0 -1.2025341 0.036457825 0.053580279762836865 0 -641 0 -1.4967335 0.020663453 0.030123371045923587 0 -642 0 -1.4967335 0.020663453 0.030123371045923587 0 -643 0 -1.0016527 0.053369574 0.079126802248061193 0 -644 0 -1.1315987 0.041740835 0.061512203112918669 0 -645 0 -1.4967335 0.020663453 0.030123371045923587 0 -646 0 -0.56830096 0.11759834 0.18049258977657712 0 -647 0 -0.91678506 0.06255057 0.093187222906957218 0 -648 1 1.0182511 0.75662625 0.40234726864350118 1 -649 0 -1.4967335 0.020663453 0.030123371045923587 0 -650 0 -0.52581793 0.12663612 0.19534522474861157 0 -651 0 -0.7802133 0.080463685 0.12102154330794489 0 -652 0 -1.4725443 0.021657925 0.031589106728466369 0 -653 0 -1.5951606 0.017058393 0.024822381042273359 0 -654 0 -1.7557883 0.012458842 0.018087218256857347 0 -655 0 -1.4967335 0.020663453 0.030123371045923587 0 -656 0 -1.6840487 0.014338439 0.020835730799420298 0 -657 0 -1.1226763 0.042455107 0.062587967715861476 0 -658 1 3.3072784 0.9965929 0.0049238289554092525 1 -659 0 -1.0016527 0.053369574 0.079126802248061193 0 -660 0 -1.263222 0.03245402 0.047597873731249518 0 -661 0 -1.8490299 0.01037536 0.015046673469226494 0 -662 0 -1.0531907 0.048430555 0.071619146823917096 0 -663 0 -1.0531907 0.048430555 0.071619146823917096 0 -664 0 -1.3577405 0.027051782 0.039565070931675883 0 -665 0 -1.0016527 0.053369574 0.079126802248061193 0 -666 0 -0.98967534 0.054583687 0.080978337026532551 0 -667 0 -1.7557883 0.012458842 0.018087218256857347 0 -668 1 0.40061146 0.47704738 1.0677955226673479 1 -669 1 3.7611184 0.9986133 0.0020019765797157029 1 -670 1 3.6400266 0.99823713 0.0025455238106964978 1 -671 0 -1.2611592 0.032582857 0.047789991346392016 0 -672 0 -1.6512232 0.015289166 0.022227963838514678 0 -673 0 -0.9058171 0.06383956 0.095172296991428637 0 -674 0 -1.263222 0.03245402 0.047597873731249518 0 -675 0 -0.9178249 0.062429626 0.09300111098076376 0 -676 0 -1.1693764 0.03884206 0.057154576473333961 0 -677 0 -1.4468354 0.022766171 0.033224288535319356 0 -678 0 -1.0016527 0.053369574 0.079126802248061193 0 -679 0 -1.1315987 0.041740835 0.061512203112918669 0 -680 1 3.730233 0.99852574 0.0021284787959572086 1 -681 1 4.8359895 0.99983567 0.00023709756587015005 1 -682 0 -1.709371 0.013645053 0.019821190416295924 0 -683 0 -1.0016527 0.053369574 0.079126802248061193 0 -684 0 -1.0016527 0.053369574 0.079126802248061193 0 -685 0 -1.0016527 0.053369574 0.079126802248061193 0 -686 0 -1.0016527 0.053369574 0.079126802248061193 0 -687 0 -1.1030984 0.044063523 0.065013342744117381 0 -688 0 -1.3522246 0.02734149 0.039994716691546361 0 -689 0 -2.6891706 0.0019739356 0.0028506015312286881 0 -690 0 -0.91678506 0.06255057 0.093187222906957218 0 -691 1 3.3331482 0.9967629 0.0046777642370595699 1 -692 0 -1.2478162 0.03342823 0.049051234404724475 0 -693 0 -1.7172247 0.013436794 0.019516612457250358 0 -694 0 -1.3637799 0.026738003 0.039099871543600485 0 -695 0 -1.1315987 0.041740835 0.061512203112918669 0 -696 1 2.3350956 0.9769887 0.03358625858753999 1 -697 1 2.0787206 0.962295 0.055448868546929708 1 -698 1 2.188851 0.9694736 0.044726481865753932 1 +618 0 -1.4413924 0.012076221 0.017528357214092197 0 +619 0 -1.5088496 0.010326769 0.014975838511155387 0 +620 0 -1.3621383 0.014508282 0.021084347628400252 0 +621 0 -0.9565075 0.03673035 0.053988382371163816 0 +622 0 -1.5543131 0.009291672 0.013467715475217655 0 +623 0 -0.9823508 0.034644295 0.050867463428077943 0 +624 0 -1.1971438 0.021221075 0.030945057450401461 0 +625 0 -0.561588 0.08785032 0.13265750658870359 0 +626 1 1.5907574 0.9375945 0.092964029048681035 1 +627 0 -0.21013032 0.18011338 0.2865036694596596 0 +628 0 -1.093269 0.02691992 0.039369557847489002 0 +629 0 -1.2739269 0.01778492 0.025889121808373532 0 +630 0 -0.8247035 0.049383637 0.073064859200447238 0 +631 0 -1.5088496 0.010326769 0.014975838511155387 0 +632 0 -0.9823508 0.034644295 0.050867463428077943 0 +633 1 1.6873877 0.9496157 0.074584281900602706 1 +634 0 -1.1901796 0.021563126 0.031449319333433555 0 +635 0 -0.96809995 0.035780076 0.052565853859723656 0 +636 1 1.918179 0.9700512 0.043867155966165396 1 +637 0 -0.5021047 0.09969573 0.15151542830038794 0 +638 0 -1.2739269 0.01778492 0.025889121808373532 0 +639 0 -1.450688 0.011818778 0.017152453443940112 0 +640 0 -1.3926399 0.013519853 0.019638078528348882 0 +641 0 -1.3621383 0.014508282 0.021084347628400252 0 +642 0 -1.3621383 0.014508282 0.021084347628400252 0 +643 0 -0.9823508 0.034644295 0.050867463428077943 0 +644 0 -1.093269 0.02691992 0.039369557847489002 0 +645 0 -1.3621383 0.014508282 0.021084347628400252 0 +646 0 -0.71524304 0.06293352 0.093776690855145756 0 +647 0 -1.1489447 0.023701586 0.034605907302734731 0 +648 1 1.8725555 0.96678 0.048740456563910198 1 +649 0 -1.3621383 0.014508282 0.021084347628400252 0 +650 0 -0.8149995 0.050463486 0.074704615959513512 0 +651 0 -1.0335566 0.030843426 0.045198332295617018 0 +652 0 -1.3365781 0.015391149 0.022377385809384916 0 +653 0 -1.4413924 0.012076221 0.017528357214092197 0 +654 0 -1.539121 0.009625588 0.013954053758422826 0 +655 0 -1.3621383 0.014508282 0.021084347628400252 0 +656 0 -1.5088496 0.010326769 0.014975838511155387 0 +657 0 -0.64531356 0.07333143 0.10987465540518587 0 +658 1 2.7625368 0.9957598 0.0061303430902180595 1 +659 0 -0.9823508 0.034644295 0.050867463428077943 0 +660 0 -1.1645964 0.022866543 0.033372476346272044 0 +661 0 -1.6143917 0.0080799395 0.01170423737885466 0 +662 0 -1.0735937 0.02815592 0.041203223287694185 0 +663 0 -1.0735937 0.02815592 0.041203223287694185 0 +664 0 -1.494573 0.0106747765 0.015483234986686177 0 +665 0 -0.9823508 0.034644295 0.050867463428077943 0 +666 0 -1.033285 0.030862475 0.045226689488849996 0 +667 0 -1.539121 0.009625588 0.013954053758422826 0 +668 1 0.9935051 0.78724366 0.34511785328654954 1 +669 1 2.9317129 0.997145 0.0041247894454964421 1 +670 1 2.6631372 0.99465203 0.007736190617521033 1 +671 0 -1.2509142 0.018753061 0.027311846580596485 0 +672 0 -1.4515353 0.011795582 0.017118589545051147 0 +673 0 -0.8807356 0.043565203 0.064261475020704048 0 +674 0 -1.1645964 0.022866543 0.033372476346272044 0 +675 0 -1.0106881 0.032488268 0.047648939278955048 0 +676 0 -1.1284822 0.024838284 0.0362866062628436 0 +677 0 -1.345466 0.015078313 0.021919076775756236 0 +678 0 -0.9823508 0.034644295 0.050867463428077943 0 +679 0 -1.093269 0.02691992 0.039369557847489002 0 +680 1 3.618086 0.9994282 0.00082515071272389008 1 +681 1 3.4351132 0.9991219 0.0012673807284635152 1 +682 0 -1.491029 0.010762947 0.01561181589588421 0 +683 0 -0.9823508 0.034644295 0.050867463428077943 0 +684 0 -0.9823508 0.034644295 0.050867463428077943 0 +685 0 -0.9823508 0.034644295 0.050867463428077943 0 +686 0 -0.9823508 0.034644295 0.050867463428077943 0 +687 0 -1.0711788 0.028311372 0.041434009989270069 0 +688 0 -1.2739269 0.01778492 0.025889121808373532 0 +689 0 -1.9011282 0.0041396352 0.0059846268911291807 0 +690 0 -1.1489447 0.023701586 0.034605907302734731 0 +691 1 2.4506643 0.99122626 0.01271369237054799 1 +692 0 -1.1901796 0.021563126 0.031449319333433555 0 +693 0 -1.4852631 0.010907938 0.015823285189527934 0 +694 0 -1.3193941 0.016014215 0.02329062026634221 0 +695 0 -1.093269 0.02691992 0.039369557847489002 0 +696 1 2.2754688 0.9868245 0.019134542609123055 1 +697 1 1.9577334 0.9726328 0.040032812989940499 1 +698 1 2.009111 0.97566503 0.035542170842136504 1 diff --git a/test/BaselineOutput/Common/LogisticRegression/netcoreapp/LogisticRegression-non-negative-CV-breast-cancer.txt b/test/BaselineOutput/Common/LogisticRegression/netcoreapp/LogisticRegression-non-negative-CV-breast-cancer.txt index 156a695123..d2fa8eb18a 100644 --- a/test/BaselineOutput/Common/LogisticRegression/netcoreapp/LogisticRegression-non-negative-CV-breast-cancer.txt +++ b/test/BaselineOutput/Common/LogisticRegression/netcoreapp/LogisticRegression-non-negative-CV-breast-cancer.txt @@ -1,700 +1,700 @@ Instance Label Score Probability Log-loss Assigned -5 1 12.634052 0.9999967 4.7295306510363196E-06 1 -6 0 -1.3831959 0.20049621 0.32282322512958106 0 -8 0 -4.7247458 0.008794933 0.012744531847526374 0 -9 0 -5.136714 0.005842633 0.0084538582850831219 0 -10 0 -5.6094384 0.0036497563 0.0052751176456810279 0 -11 0 -5.869679 0.0028158284 0.004068111950303187 0 -18 1 7.2676487 0.99930274 0.0010062773817144412 1 -20 1 7.5638084 0.9994814 0.00074840460816191399 1 -21 1 7.3895473 0.99938273 0.00089080112495110294 1 -25 1 1.3491688 0.7939937 0.33280051379712122 1 -28 0 -5.869679 0.0028158284 0.004068111950303187 0 -31 0 -5.1811585 0.0055900654 0.0080873852307543487 0 -32 1 7.2914295 0.99931914 0.00098261346137125155 1 -35 0 -5.869679 0.0028158284 0.004068111950303187 0 -37 0 -1.0245552 0.26414105 0.44249884478738094 0 +5 1 12.6340475 0.9999967 4.7295306510363196E-06 1 +6 0 -1.383316 0.20047696 0.32278848500188356 0 +8 0 -4.7247133 0.008795215 0.012744942575086885 0 +9 0 -5.1366987 0.0058427216 0.0084539866786188071 0 +10 0 -5.60942 0.003649824 0.0052752157516993657 0 +11 0 -5.8696694 0.0028158552 0.0040681506882989755 0 +18 1 7.267564 0.9993027 0.0010063634330421671 1 +20 1 7.56392 0.99948144 0.00074831857221908471 1 +21 1 7.389552 0.99938273 0.00089080112495110294 1 +25 1 1.3491125 0.7939845 0.33281730074743887 1 +28 0 -5.8696694 0.0028158552 0.0040681506882989755 0 +31 0 -5.1811304 0.005590222 0.0080876122269247976 0 +32 1 7.2914686 0.99931914 0.00098261346137125155 1 +35 0 -5.8696694 0.0028158552 0.0040681506882989755 0 +37 0 -1.024457 0.26416016 0.44253629840217606 0 40 0 ? ? ? 0 -41 1 2.7845802 0.94183683 0.086450949311470543 1 -44 1 7.8202686 0.9995986 0.00057918182856106851 1 -45 0 -5.7772603 0.0030876263 0.0044613942665847131 0 -46 1 4.3640785 0.98743355 0.018244426563395755 1 -48 0 -3.8105168 0.021657312 0.031588203058819132 0 -50 1 2.3512 0.91302955 0.13126653903574156 1 -51 1 -0.38567162 0.4047597 1.3048624219352793 0 -52 1 4.762245 0.991526 0.012277480154992908 1 -54 1 7.3692417 0.99937004 0.00090912870659733396 1 -56 1 4.515278 0.9891778 0.015698199387500632 1 -60 1 2.0932302 0.8902435 0.16772814544847323 1 -63 1 -0.43255043 0.39351746 1.345500430582274 0 -64 0 -6.235624 0.0019545793 0.0028226212236700603 0 -66 0 -4.449269 0.011552098 0.016763167178581874 0 -68 1 12.420467 0.99999595 5.8474219794828371E-06 1 -69 0 -5.3731294 0.004618163 0.0066780331962603507 0 -70 0 -3.9443603 0.018995773 0.027668742142448771 0 -71 1 8.000187 0.9996647 0.00048378231041388311 1 -72 0 -1.4700346 0.18693736 0.2985615936142022 0 -73 1 8.232616 0.9997342 0.00038348628329455832 1 -74 1 1.5340109 0.8225924 0.28175039134378066 1 -76 0 -4.994397 0.006730202 0.0097424504512254439 0 -77 0 -4.4503174 0.011540132 0.01674570141466946 0 -79 0 -5.8073435 0.002996398 0.004329377909008233 0 -82 0 -4.082248 0.016589638 0.024134538030866414 0 -88 0 -4.449269 0.011552098 0.016763167178581874 0 -90 0 -5.547103 0.0038835946 0.0056137505733873209 0 -91 0 -5.872879 0.0028068572 0.0040551327591589045 0 -92 0 -4.449269 0.011552098 0.016763167178581874 0 -93 0 -6.235624 0.0019545793 0.0028226212236700603 0 -95 0 -5.547103 0.0038835946 0.0056137505733873209 0 -96 0 -6.238824 0.0019483466 0.002813611838319385 0 -97 0 -4.083324 0.016572097 0.024108805732293438 0 -98 1 8.655651 0.9998259 0.00025120253004900898 1 -99 1 11.467838 0.9999895 1.5134552659953847E-05 1 -100 1 3.7769547 0.97762 0.032654286719703295 1 -102 0 -4.0355268 0.01736934 0.025278839074877907 0 -104 1 12.721792 0.999997 4.2995726784359523E-06 1 -105 1 2.7618742 0.9405805 0.088376691713166353 1 -106 1 8.5678215 0.9998099 0.00027425240246685845 1 -108 0 -5.8876767 0.0027657421 0.0039956504761246604 0 -109 1 7.138489 0.99920666 0.0011449987869914802 1 -111 1 2.788124 0.94203067 0.08615406665814028 1 -112 1 9.633083 0.9999345 9.4507562065679747E-05 1 -113 1 9.298395 0.99990845 0.0001320887224967665 1 -115 0 -5.073519 0.0062214015 0.0090036218037672956 0 -117 1 11.031485 0.9999838 2.3389830121236442E-05 1 -120 0 -4.813985 0.008050126 0.011660875880859801 0 -121 0 -3.985537 0.018243454 0.026562783174295196 0 -122 1 9.103445 0.9998887 0.00016055473816202313 1 -123 1 3.8367043 0.97889066 0.030780375815034355 1 -125 0 -6.235624 0.0019545793 0.0028226212236700603 0 -128 1 3.931861 0.9807699 0.028013431951430235 1 -129 0 -6.425461 0.0016171673 0.0023349677143088441 0 -131 0 -5.1811585 0.0055900654 0.0080873852307543487 0 -132 1 8.02145 0.99967176 0.00047363196654040616 1 -133 0 -5.3297606 0.004821867 0.0069733096511885159 0 -137 0 -6.192255 0.0020410353 0.0029476006542974429 0 -138 0 -4.724047 0.008801024 0.012753398437778941 0 -141 0 -6.5582 0.0014164277 0.0020449217526388465 0 -144 0 -5.869679 0.0028158284 0.004068111950303187 0 +41 1 2.7846231 0.9418392 0.086447297247185473 1 +44 1 7.8203325 0.9995987 0.00057909580270934437 1 +45 0 -5.777241 0.003087685 0.0044614791764681647 0 +46 1 4.3639555 0.987432 0.018246690792932486 1 +48 0 -3.8104544 0.021658637 0.031590155974785178 0 +50 1 2.3510952 0.91302127 0.13127963045041355 1 +51 1 -0.38567448 0.40475902 1.3048648651159307 0 +52 1 4.76217 0.9915254 0.012278347417683086 1 +54 1 7.369298 0.9993701 0.00090904266106909778 1 +56 1 4.5150814 0.98917574 0.015701242014916676 1 +60 1 2.0932417 0.89024454 0.16772640677475523 1 +63 1 -0.43256855 0.39351314 1.3455162733488315 0 +64 0 -6.2356153 0.001954596 0.0028226454560947357 0 +66 0 -4.449239 0.011552442 0.016763668766800108 0 +68 1 12.420444 0.99999595 5.8474219794828371E-06 1 +69 0 -5.373103 0.004618284 0.0066782080016231423 0 +70 0 -3.9443884 0.01899525 0.027667972409654292 0 +71 1 8.000552 0.99966484 0.00048361027009226202 1 +72 0 -1.4699545 0.18694954 0.29858319571334258 0 +73 1 8.232631 0.9997342 0.00038348628329455832 1 +74 1 1.5339813 0.8225881 0.28175791802619315 1 +76 0 -4.9944496 0.0067298515 0.0097419411527891794 0 +77 0 -4.450281 0.011540545 0.016746304944427276 0 +79 0 -5.8073263 0.0029964494 0.0043294523668142692 0 +82 0 -4.0822144 0.016590191 0.024135349601747052 0 +88 0 -4.449239 0.011552442 0.016763668766800108 0 +90 0 -5.547076 0.003883698 0.0056139002960622839 0 +91 0 -5.87287 0.0028068826 0.0040551694757115222 0 +92 0 -4.449239 0.011552442 0.016763668766800108 0 +93 0 -6.2356153 0.001954596 0.0028226454560947357 0 +95 0 -5.547076 0.003883698 0.0056139002960622839 0 +96 0 -6.238816 0.0019483624 0.0028136347243553146 0 +97 0 -4.083293 0.016572602 0.024109546243338689 0 +98 1 8.65561 0.9998259 0.00025120253004900898 1 +99 1 11.467908 0.99998957 1.5048560434990871E-05 1 +100 1 3.7769213 0.9776193 0.032655342238490488 1 +102 0 -4.035517 0.017369503 0.025279079731081816 0 +104 1 12.722043 0.999997 4.2995726784359523E-06 1 +105 1 2.761939 0.94058406 0.088371206303052993 1 +106 1 8.567689 0.99980986 0.00027433841014312443 1 +108 0 -5.8876686 0.0027657645 0.0039956828123055965 0 +109 1 7.1384897 0.99920666 0.0011449987869914802 1 +111 1 2.7880402 0.9420261 0.086161095461353254 1 +112 1 9.63318 0.9999345 9.4507562065679747E-05 1 +113 1 9.298348 0.99990845 0.0001320887224967665 1 +115 0 -5.0735936 0.006220942 0.0090029545791122101 0 +117 1 11.031608 0.99998385 2.3303837404214046E-05 1 +120 0 -4.8139625 0.008050305 0.01166113594843611 0 +121 0 -3.98552 0.018243762 0.026563234806467287 0 +122 1 9.103314 0.9998887 0.00016055473816202313 1 +123 1 3.8365965 0.97888845 0.030783626109233086 1 +125 0 -6.2356153 0.001954596 0.0028226454560947357 0 +128 1 3.9317198 0.98076725 0.028017289760976786 1 +129 0 -6.4254074 0.0016172534 0.0023350921999658724 0 +131 0 -5.1811304 0.005590222 0.0080876122269247976 0 +132 1 8.021436 0.99967176 0.00047363196654040616 1 +133 0 -5.3297505 0.004821915 0.006973379182606129 0 +137 0 -6.1922626 0.0020410197 0.0029475781027268637 0 +138 0 -4.7240562 0.008800945 0.012753283216489029 0 +141 0 -6.5582085 0.0014164156 0.0020449042608751558 0 +144 0 -5.8696694 0.0028158552 0.0040681506882989755 0 145 0 ? ? ? 0 147 0 -5.710632 0.0032996563 0.0047682689732922723 0 -150 0 -5.6094384 0.0036497563 0.0052751176456810279 0 -151 1 5.1198063 0.9940583 0.0085976128158061754 1 -152 1 9.2328205 0.99990225 0.00014103266690546063 1 -154 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -156 0 -5.8208055 0.002956449 0.0042715718473290056 0 -161 0 -4.2752953 0.013717164 0.019926668193276719 0 +150 0 -5.60942 0.003649824 0.0052752157516993657 0 +151 1 5.1197214 0.99405783 0.0085983048584769572 1 +152 1 9.2328615 0.99990225 0.00014103266690546063 1 +154 0 -6.924154 0.0009827684 0.0014185322093419092 0 +156 0 -5.820814 0.0029564237 0.0042715351252685394 0 +161 0 -4.2752657 0.013717564 0.019927252620683504 0 164 0 ? ? ? 0 -167 1 8.726606 0.9998378 0.00023400137267902865 1 -169 0 -6.6671042 0.0012704609 0.0018340529276932821 0 -171 0 -5.547103 0.0038835946 0.0056137505733873209 0 -173 1 16.563875 0.99999994 8.5991327994145617E-08 1 -174 1 5.9247894 0.9973348 0.0038502352641274578 1 -176 0 -5.1811585 0.0055900654 0.0080873852307543487 0 -177 1 4.522333 0.9892531 0.015588408297398782 1 -179 1 1.7983112 0.85794324 0.22104589547952272 1 -180 0 -5.6094384 0.0036497563 0.0052751176456810279 0 -181 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -183 1 9.378843 0.99991554 0.00012185485412592618 1 -187 1 15.030241 0.9999997 4.2995669122556443E-07 1 -188 1 9.162302 0.9998951 0.00015135267166164302 1 -189 0 -4.7185426 0.008849174 0.012823481797966271 0 -191 1 11.084532 0.9999847 2.2099939904093242E-05 1 -192 0 -4.771845 0.008393699 0.012160656505235928 0 -196 0 6.867338 0.99895984 9.9089779104508953 1 -198 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -199 0 -5.503734 0.0040550292 0.0058620640547876936 0 -201 1 9.7134495 0.99993956 8.7197839093874189E-05 1 -202 0 -5.547103 0.0038835946 0.0056137505733873209 0 -204 0 -5.547103 0.0038835946 0.0056137505733873209 0 -205 1 12.771593 0.99999714 4.1275895252744015E-06 1 -206 1 4.5629673 0.9896766 0.014970934743901675 1 -207 0 -5.6094384 0.0036497563 0.0052751176456810279 0 -209 0 -4.14566 0.015586208 0.022663223704780491 0 -210 1 15.273354 0.99999976 3.4396534272948301E-07 1 -211 1 9.317207 0.9999101 0.00012968074693731126 1 -212 0 -5.547103 0.0038835946 0.0056137505733873209 0 -216 0 -6.235624 0.0019545793 0.0028226212236700603 0 -218 1 7.7950516 0.99958843 0.00059389232464231211 1 -219 0 -2.889895 0.052655358 0.078038724059753192 0 -223 1 5.9372654 0.99736774 0.0038025557705296848 1 -226 1 10.314253 0.99996686 4.7811969192929004E-05 1 -228 0 -5.6094384 0.0036497563 0.0052751176456810279 0 -233 1 5.7215347 0.996736 0.0047166727998034655 1 -237 1 5.903057 0.99727637 0.0039347344400170363 1 -239 1 5.306081 0.9950632 0.0071399563768214189 1 -240 0 -2.4950051 0.07620908 0.11436173396919315 0 -241 0 -4.2387977 0.014219806 0.020662099195145979 0 -242 0 -5.1811585 0.0055900654 0.0080873852307543487 0 -244 0 -5.547103 0.0038835946 0.0056137505733873209 0 -246 1 11.306314 0.9999877 1.7714321792245208E-05 1 -247 1 2.3081617 0.9095507 0.13677399444352431 1 -248 0 -3.5794015 0.027135514 0.039689233962837453 0 +167 1 8.726754 0.9998378 0.00023400137267902865 1 +169 0 -6.6671047 0.0012704603 0.0018340520868660067 0 +171 0 -5.547076 0.003883698 0.0056139002960622839 0 +173 1 16.563984 0.99999994 8.5991327994145617E-08 1 +174 1 5.9247246 0.9973346 0.0038504939275221689 1 +176 0 -5.1811304 0.005590222 0.0080876122269247976 0 +177 1 4.5223436 0.9892532 0.015588234446399627 1 +179 1 1.7982273 0.857933 0.22106313507733019 1 +180 0 -5.60942 0.003649824 0.0052752157516993657 0 +181 0 -6.924154 0.0009827684 0.0014185322093419092 0 +183 1 9.378819 0.99991554 0.00012185485412592618 1 +187 1 15.030385 0.9999997 4.2995669122556443E-07 1 +188 1 9.162457 0.9998951 0.00015135267166164302 1 +189 0 -4.7185535 0.008849078 0.012823342170092279 0 +191 1 11.084495 0.9999846 2.2185932549356626E-05 1 +192 0 -4.771832 0.008393806 0.012160812328843666 0 +196 0 6.867385 0.9989599 9.9090605840156236 1 +198 0 -6.924154 0.0009827684 0.0014185322093419092 0 +199 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +201 1 9.71342 0.99993956 8.7197839093874189E-05 1 +202 0 -5.547076 0.003883698 0.0056139002960622839 0 +204 0 -5.547076 0.003883698 0.0056139002960622839 0 +205 1 12.771596 0.99999714 4.1275895252744015E-06 1 +206 1 4.5628767 0.9896757 0.014972238069120203 1 +207 0 -5.60942 0.003649824 0.0052752157516993657 0 +209 0 -4.1456366 0.015586566 0.022663749186712233 0 +210 1 15.2733965 0.99999976 3.4396534272948301E-07 1 +211 1 9.317216 0.9999101 0.00012968074693731126 1 +212 0 -5.547076 0.003883698 0.0056139002960622839 0 +216 0 -6.2356153 0.001954596 0.0028226454560947357 0 +218 1 7.795169 0.99958843 0.00059389232464231211 1 +219 0 -2.889904 0.052654907 0.078038037604988009 0 +223 1 5.9372063 0.99736756 0.0038028144253759934 1 +226 1 10.314187 0.99996686 4.7811969192929004E-05 1 +228 0 -5.60942 0.003649824 0.0052752157516993657 0 +233 1 5.7214584 0.99673575 0.0047170178915291556 1 +237 1 5.903097 0.9972765 0.0039345619876793607 1 +239 1 5.3060102 0.9950628 0.0071404748846435589 1 +240 0 -2.495017 0.07620824 0.11436041914037622 0 +241 0 -4.238744 0.014220561 0.020663204585323822 0 +242 0 -5.1811304 0.005590222 0.0080876122269247976 0 +244 0 -5.547076 0.003883698 0.0056139002960622839 0 +246 1 11.306256 0.9999877 1.7714321792245208E-05 1 +247 1 2.308014 0.90953857 0.1367932812706758 1 +248 0 -3.5793734 0.027136255 0.039690333311752569 0 249 0 ? ? ? 0 -250 0 -6.5093265 0.0014872673 0.0021472703657269988 0 -252 0 4.2286224 0.98563683 6.121482030952655 1 -254 1 5.1990414 0.99450845 0.0079444725591774987 1 -257 0 -5.503734 0.0040550292 0.0058620640547876936 0 -258 0 -4.8152137 0.008040319 0.011646612871512004 0 -259 0 2.3531055 0.91318077 3.5258415372989349 1 -260 1 10.041292 0.9999564 6.2861028359088115E-05 1 -262 1 9.612917 0.9999331 9.6485493471777027E-05 1 -267 1 4.141609 0.9843515 0.022754495279888837 1 -268 1 10.114798 0.9999595 5.8389291528570374E-05 1 -269 0 -5.547103 0.0038835946 0.0056137505733873209 0 -271 0 -4.083324 0.016572097 0.024108805732293438 0 -272 1 4.141609 0.9843515 0.022754495279888837 1 +250 0 -6.509353 0.0014872277 0.002147213177058749 0 +252 0 4.228568 0.9856361 6.1214101895529778 1 +254 1 5.1988707 0.99450755 0.0079457695521466531 1 +257 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +258 0 -4.8151846 0.008040551 0.011646950143324645 0 +259 0 2.3529425 0.91316783 3.5256266226141006 1 +260 1 10.041265 0.9999564 6.2861028359088115E-05 1 +262 1 9.612809 0.9999331 9.6485493471777027E-05 1 +267 1 4.141512 0.98435 0.022756679240322205 1 +268 1 10.114848 0.9999595 5.8389291528570374E-05 1 +269 0 -5.547076 0.003883698 0.0056139002960622839 0 +271 0 -4.083293 0.016572602 0.024109546243338689 0 +272 1 4.141512 0.98435 0.022756679240322205 1 275 0 ? ? ? 0 -276 0 -5.503734 0.0040550292 0.0058620640547876936 0 -277 0 -6.235624 0.0019545793 0.0028226212236700603 0 -278 0 -5.547103 0.0038835946 0.0056137505733873209 0 -279 1 7.266674 0.999302 0.0010073099979859077 1 -280 0 -4.8152137 0.008040319 0.011646612871512004 0 -283 1 5.7072334 0.99668914 0.0047844849094791442 1 -284 1 7.021229 0.9991081 0.0012873483875371966 1 -285 1 14.94039 0.9999997 4.2995669122556443E-07 1 -288 1 2.028204 0.88372666 0.17832789312655251 1 -290 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -291 0 -5.547103 0.0038835946 0.0056137505733873209 0 -293 1 4.411065 0.98800343 0.01741204052294499 1 -296 0 1.506464 0.8185366 2.4622493375152015 1 +276 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +277 0 -6.2356153 0.001954596 0.0028226454560947357 0 +278 0 -5.547076 0.003883698 0.0056139002960622839 0 +279 1 7.266614 0.999302 0.0010073099979859077 1 +280 0 -4.8151846 0.008040551 0.011646950143324645 0 +283 1 5.7070665 0.9966886 0.0047852614024752813 1 +284 1 7.021199 0.999108 0.0012874344556313836 1 +285 1 14.940461 0.9999997 4.2995669122556443E-07 1 +288 1 2.0281782 0.88372403 0.17833217456799788 1 +290 0 -6.924154 0.0009827684 0.0014185322093419092 0 +291 0 -5.547076 0.003883698 0.0056139002960622839 0 +293 1 4.410968 0.9880023 0.017413694197482368 1 +296 0 1.5064402 0.81853306 2.4622213790482594 1 297 0 ? ? ? 0 -299 1 7.9359417 0.9996425 0.00051586818910542408 1 -300 1 7.851406 0.99961096 0.00056137458662143525 1 -301 0 -5.547103 0.0038835946 0.0056137505733873209 0 -303 0 -5.547103 0.0038835946 0.0056137505733873209 0 -304 1 5.387084 0.99544555 0.0065856916808891101 1 -308 1 7.820923 0.9995989 0.00057875169935374328 1 +299 1 7.9359903 0.9996425 0.00051586818910542408 1 +300 1 7.851367 0.99961096 0.00056137458662143525 1 +301 0 -5.547076 0.003883698 0.0056139002960622839 0 +303 0 -5.547076 0.003883698 0.0056139002960622839 0 +304 1 5.3870573 0.99544543 0.0065858644504205519 1 +308 1 7.8210382 0.999599 0.00057866567352766714 1 309 0 -1.5875473 0.16972925 0.26834621707910555 0 -311 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -312 1 2.3300753 0.91133744 0.13394276422007667 1 -314 0 -6.2979593 0.0018366763 0.002652200104836864 0 -316 1 3.2827053 0.9638307 0.053148326102386338 1 -317 1 9.330563 0.9999113 0.00012796076685551788 1 -319 0 2.9553108 0.9505139 4.3368328222592689 1 +311 0 -6.924154 0.0009827684 0.0014185322093419092 0 +312 1 2.3302755 0.9113536 0.13391719362265342 1 +314 0 -6.2979584 0.001836678 0.0026522026287495886 0 +316 1 3.2825785 0.9638263 0.053154928270775177 1 +317 1 9.330617 0.99991137 0.00012787476790525535 1 +319 0 2.9553766 0.950517 4.3369231847832852 1 321 0 ? ? ? 0 -323 1 5.7154255 0.9967161 0.0047454882432208332 1 -327 0 -6.235624 0.0019545793 0.0028226212236700603 0 -328 1 1.3741245 0.7980457 0.32545673964974264 1 -329 1 7.881464 0.99962246 0.00054477189720687899 1 -331 0 -3.6249604 0.025958357 0.037944642580990014 0 -332 0 -3.8636775 0.020559115 0.029969675365243253 0 -333 1 4.9105396 0.9926854 0.010591553470219487 1 -336 1 4.988022 0.99322706 0.0098045205997943641 1 -338 0 -6.2979593 0.0018366763 0.002652200104836864 0 -343 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -344 1 10.53714 0.9999735 3.8266647312775575E-05 1 -346 0 -4.3158894 0.013178671 0.019139196406807307 0 -347 0 -6.638533 0.0013072349 0.0018871750186523046 0 -348 1 -0.16142368 0.4597315 1.1211366076012741 0 -349 1 4.5816746 0.989866 0.014694830121292021 1 -350 0 -4.8216143 0.007989431 0.011572603255977185 0 -352 0 0.27711296 0.5688383 1.213699059773095 1 -353 1 9.354343 0.9999134 0.00012495080664652852 1 -354 0 -6.235624 0.0019545793 0.0028226212236700603 0 -355 0 -4.9014807 0.007380686 0.010687568245319782 0 -358 1 4.250904 0.9859489 0.020415189066112244 1 -360 1 16.686596 0.99999994 8.5991327994145617E-08 1 -361 1 6.061717 0.997675 0.0033581690143690331 1 -366 1 14.498312 0.9999995 6.8793076746672365E-07 1 -368 0 -6.09666 0.0022453184 0.0032429517818591126 0 -370 0 -4.65956 0.009381776 0.013598932625378389 0 -371 0 -6.09666 0.0022453184 0.0032429517818591126 0 -373 0 -4.4524693 0.01151561 0.016709911464012926 0 -376 0 -6.235624 0.0019545793 0.0028226212236700603 0 -377 0 -6.2979593 0.0018366763 0.002652200104836864 0 -378 0 -4.309298 0.013264666 0.019264924064685877 0 -379 0 -2.7152743 0.062078048 0.092460218510323433 0 -381 1 9.81171 0.9999452 7.9028192541775766E-05 1 -383 0 -6.5582 0.0014164277 0.0020449217526388465 0 -384 0 -6.5582 0.0014164277 0.0020449217526388465 0 -387 0 -2.8719764 0.053556386 0.079411538008570717 0 -388 0 -5.6957054 0.0033491116 0.0048398559067134002 0 -389 0 -4.07288 0.016743172 0.024359795163288241 0 -391 1 9.566624 0.99992996 0.00010104334573168113 1 -392 0 -5.503734 0.0040550292 0.0058620640547876936 0 -395 0 -5.503734 0.0040550292 0.0058620640547876936 0 -396 0 -4.8152137 0.008040319 0.011646612871512004 0 -398 0 -5.328685 0.004827032 0.0069807967864821494 0 -399 0 -6.095584 0.0022477307 0.0032464399153914179 0 -404 0 -6.1103435 0.0022148718 0.0031989284977273821 0 -406 0 -4.7229714 0.008810414 0.012767065103584952 0 -409 0 -5.089992 0.006120379 0.0088569722979391138 0 -413 0 -3.669582 0.024853675 0.036309376586392543 0 -414 1 6.568573 0.99859816 0.0020238488722975523 1 -415 0 -1.1188984 0.24621567 0.40777629344186561 0 -416 1 8.785311 0.99984705 0.00022067061677813278 1 -418 0 -2.254137 0.09499321 0.14399948065253987 0 -419 0 -6.3690414 0.0017108691 0.0024703762708912901 0 -422 0 -3.2716856 0.036555417 0.053726409124214863 0 -423 0 -3.9443603 0.018995773 0.027668742142448771 0 -428 0 -6.235624 0.0019545793 0.0028226212236700603 0 -429 0 -5.869679 0.0028158284 0.004068111950303187 0 -430 0 -6.059499 0.002330129 0.0033655882551841031 0 -434 0 4.339279 0.98712206 6.2789542296040644 1 -436 1 2.258398 0.9053725 0.14341660773032266 1 -439 0 -5.460366 0.0042339973 0.0061213349516453406 0 -440 1 10.559719 0.9999741 3.7406711505747454E-05 1 -441 0 -2.6994514 0.06300573 0.093887870272153742 0 -442 0 -5.3600802 0.004678537 0.0067655411147107191 0 -449 1 10.890316 0.99998134 2.691553593234519E-05 1 -450 0 -4.561373 0.010339676 0.014994654089518316 0 -451 0 -5.460366 0.0042339973 0.0061213349516453406 0 -452 0 -5.5535035 0.0038589125 0.0055780033806187143 0 -453 1 9.72237 0.9999401 8.6423870594618947E-05 1 -454 0 -6.4937124 0.0015106365 0.0021810356240617268 0 -455 1 -0.44417953 0.39074552 1.3556987602603254 0 -456 1 10.305763 0.99996656 4.8241940133472131E-05 1 -457 1 9.48892 0.9999243 0.00010921311695177715 1 -464 0 -5.82631 0.0029402673 0.0042481575122417245 0 -465 1 9.441341 0.9999206 0.00011454499259225444 1 -466 1 9.603409 0.9999325 9.7345464494158371E-05 1 -467 1 8.005106 0.99966633 0.00048145976780300705 1 -474 0 -5.460366 0.0042339973 0.0061213349516453406 0 -480 0 -5.5227013 0.0039791446 0.0057521441187991726 0 -482 1 16.848186 0.99999994 8.5991327994145617E-08 1 -483 1 11.242464 0.9999869 1.8918215632667518E-05 1 -484 0 -5.0466237 0.0063899164 0.009248280152246215 0 -487 1 14.043451 0.9999992 1.1178876637117433E-06 1 -489 1 -1.1626606 0.23818418 2.069850477902655 0 -492 0 -5.200125 0.0054856157 0.0079358570427709406 0 -493 1 10.15749 0.9999612 5.5981438976225112E-05 1 -495 0 -5.5660696 0.003810908 0.005508480832347523 0 -497 0 -5.8252344 0.0029434226 0.0042527231093105935 0 -501 0 -5.1377897 0.0058363876 0.00844479508184608 0 -502 0 -4.8184137 0.008014836 0.011609551378642757 0 -504 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -507 0 -5.5361757 0.0039260965 0.005675308135661126 0 -510 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -513 0 -5.5660696 0.003810908 0.005508480832347523 0 -514 1 11.337807 0.9999881 1.7198367596743667E-05 1 -517 0 -6.2979593 0.0018366763 0.002652200104836864 0 -519 1 7.511837 0.9994537 0.00078832583915587986 1 -520 0 -6.861809 0.0010459233 0.0015097380753681191 0 -521 0 -5.763975 0.0031287903 0.0045209665168748185 0 -522 1 5.377714 0.9954029 0.0066475444953338209 1 -523 1 7.6095295 0.99950457 0.0007149370135965138 1 -527 0 -4.449269 0.011552098 0.016763167178581874 0 -528 0 -3.9475608 0.018936224 0.027581170551175441 0 -529 0 -5.200125 0.0054856157 0.0079358570427709406 0 -531 0 -4.7229714 0.008810414 0.012767065103584952 0 -532 0 -5.6094384 0.0036497563 0.0052751176456810279 0 -533 0 -5.503734 0.0040550292 0.0058620640547876936 0 -534 0 -5.869679 0.0028158284 0.004068111950303187 0 -535 0 -4.400243 0.012125527 0.017600360813711864 0 -538 0 -5.1377897 0.0058363876 0.00844479508184608 0 -539 0 -4.4059005 0.012057942 0.01750166398369097 0 -540 0 -4.309189 0.013266095 0.019267012878173298 0 -541 0 -6.192255 0.0020410353 0.0029476006542974429 0 -544 0 -4.722819 0.008811746 0.012769003552087191 0 -546 1 11.498789 0.99998987 1.4618599387059818E-05 1 -547 0 -6.6205354 0.0013309434 0.0019214244220232071 0 -548 0 -6.2545905 0.0019179272 0.0027696408611504681 0 -549 1 5.805393 0.9969978 0.0043378118803141354 1 -557 0 -4.8216143 0.007989431 0.011572603255977185 0 -558 0 -5.869679 0.0028158284 0.004068111950303187 0 -559 0 -4.771845 0.008393699 0.012160656505235928 0 -560 0 -4.083324 0.016572097 0.024108805732293438 0 -561 0 -4.083324 0.016572097 0.024108805732293438 0 -563 0 -5.503734 0.0040550292 0.0058620640547876936 0 -565 1 13.257515 0.9999983 2.4937505927785606E-06 1 -566 0 -4.4014716 0.012110815 0.017578876731927252 0 -569 1 10.79422 0.9999795 2.9581319217091579E-05 1 -577 0 -6.235624 0.0019545793 0.0028226212236700603 0 -578 0 -6.235624 0.0019545793 0.0028226212236700603 0 -581 1 9.044582 0.999882 0.00017027287211533984 1 -582 1 7.144678 0.99921155 0.0011379419170697886 1 -584 0 -3.8011508 0.021856654 0.03188218923845184 0 -586 1 14.420912 0.99999946 7.7392213646475052E-07 1 -590 1 4.6784306 0.990792 0.013345909367238451 1 -593 0 -5.0466237 0.0063899164 0.009248280152246215 0 -594 1 5.308386 0.99507445 0.0071236234758126496 1 -600 0 -5.503734 0.0040550292 0.0058620640547876936 0 -602 0 -5.1377897 0.0058363876 0.00844479508184608 0 -604 1 6.8200817 0.99890953 0.0015740696903399755 1 -606 0 -5.506934 0.0040421262 0.005843373277187037 0 -607 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -609 0 -5.460366 0.0042339973 0.0061213349516453406 0 -612 1 18.546185 1 -0 1 -613 0 -5.8675275 0.002821876 0.0040768613741309339 0 -614 0 -5.9320145 0.0026461114 0.0038225915547429492 0 +323 1 5.7153835 0.99671596 0.0047456607925197127 1 +327 0 -6.2356153 0.001954596 0.0028226454560947357 0 +328 1 1.3741446 0.798049 0.32545081328084507 1 +329 1 7.881404 0.99962246 0.00054477189720687899 1 +331 0 -3.6249194 0.025959395 0.037946179257903952 0 +332 0 -3.8636656 0.020559356 0.029970029294284398 0 +333 1 4.910447 0.9926847 0.010592506345031052 1 +336 1 4.9879694 0.9932267 0.0098050400661519183 1 +338 0 -6.2979584 0.001836678 0.0026522026287495886 0 +343 0 -6.924154 0.0009827684 0.0014185322093419092 0 +344 1 10.537286 0.9999735 3.8266647312775575E-05 1 +346 0 -4.315888 0.013178689 0.019139223637967252 0 +347 0 -6.6385508 0.0013072118 0.0018871417206665186 0 +348 1 -0.16127491 0.45976844 1.1210206432170668 0 +349 1 4.5816984 0.98986626 0.014694482634608021 1 +350 0 -4.821586 0.007989653 0.011572926966128978 0 +352 0 0.27712536 0.56884134 1.2137092313004822 1 +353 1 9.354295 0.9999134 0.00012495080664652852 1 +354 0 -6.2356153 0.001954596 0.0028226454560947357 0 +355 0 -4.9014616 0.0073808255 0.010687771286083015 0 +358 1 4.2509537 0.9859496 0.020414229681456703 1 +360 1 16.686714 0.99999994 8.5991327994145617E-08 1 +361 1 6.0618353 0.9976753 0.0033577380558281853 1 +366 1 14.498368 0.9999995 6.8793076746672365E-07 1 +368 0 -6.0967107 0.002245205 0.0032427878286808349 0 +370 0 -4.6595545 0.009381829 0.013599009936722344 0 +371 0 -6.0967107 0.002245205 0.0032427878286808349 0 +373 0 -4.4524393 0.011515952 0.016710410315180405 0 +376 0 -6.2356153 0.001954596 0.0028226454560947357 0 +377 0 -6.2979584 0.001836678 0.0026522026287495886 0 +378 0 -4.3092303 0.013265552 0.019266219019757586 0 +379 0 -2.7152834 0.06207752 0.092459404825564293 0 +381 1 9.811772 0.9999452 7.9028192541775766E-05 1 +383 0 -6.5582085 0.0014164156 0.0020449042608751558 0 +384 0 -6.5582085 0.0014164156 0.0020449042608751558 0 +387 0 -2.8719559 0.053557426 0.079413122333869088 0 +388 0 -5.6956964 0.003349142 0.0048398997209226869 0 +389 0 -4.072876 0.016743235 0.024359888084876588 0 +391 1 9.566541 0.99992996 0.00010104334573168113 1 +392 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +395 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +396 0 -4.8151846 0.008040551 0.011646950143324645 0 +398 0 -5.3286715 0.004827096 0.006980889945564733 0 +399 0 -6.0956316 0.0022476236 0.0032462850516448241 0 +404 0 -6.110317 0.0022149298 0.003199012323393043 0 +406 0 -4.7229767 0.008810367 0.012766997325712421 0 +409 0 -5.090002 0.006120318 0.0088568837492431062 0 +413 0 -3.6695714 0.02485393 0.03630975411989764 0 +414 1 6.568448 0.998598 0.0020241072084433102 1 +415 0 -1.1188269 0.24622895 0.40780170487018003 0 +416 1 8.785275 0.99984705 0.00022067061677813278 1 +418 0 -2.2541132 0.09499526 0.14400274687683212 0 +419 0 -6.36902 0.0017109058 0.0024704292663799651 0 +422 0 -3.2716913 0.036555216 0.053726107891847036 0 +423 0 -3.9443884 0.01899525 0.027667972409654292 0 +428 0 -6.2356153 0.001954596 0.0028226454560947357 0 +429 0 -5.8696694 0.0028158552 0.0040681506882989755 0 +430 0 -6.0594835 0.0023301644 0.0033656394317825608 0 +434 0 4.3392076 0.98712116 6.2788540718929182 1 +436 1 2.2582903 0.90536326 0.14343132954199816 1 +439 0 -5.460371 0.004233975 0.0061213025677854908 0 +440 1 10.559865 0.9999741 3.7406711505747454E-05 1 +441 0 -2.6994724 0.06300449 0.093885965971473181 0 +442 0 -5.3600893 0.004678495 0.0067654803678548767 0 +449 1 10.890288 0.99998134 2.691553593234519E-05 1 +450 0 -4.5613365 0.010340052 0.014995201223435974 0 +451 0 -5.460371 0.004233975 0.0061213025677854908 0 +452 0 -5.553478 0.0038590115 0.0055781466926909133 0 +453 1 9.722455 0.9999401 8.6423870594618947E-05 1 +454 0 -6.4937067 0.0015106451 0.0021810480712987851 0 +455 1 -0.44426727 0.39072463 1.3557758968203855 0 +456 1 10.305816 0.99996656 4.8241940133472131E-05 1 +457 1 9.489018 0.9999243 0.00010921311695177715 1 +464 0 -5.826317 0.0029402478 0.0042481292131313943 0 +465 1 9.44143 0.99992067 0.00011445899444170153 1 +466 1 9.603353 0.9999325 9.7345464494158371E-05 1 +467 1 8.005069 0.99966633 0.00048145976780300705 1 +474 0 -5.460371 0.004233975 0.0061213025677854908 0 +480 0 -5.522714 0.003979094 0.0057520705992681288 0 +482 1 16.84843 0.99999994 8.5991327994145617E-08 1 +483 1 11.242423 0.9999869 1.8918215632667518E-05 1 +484 0 -5.046649 0.0063897558 0.0092480468882330819 0 +487 1 14.043496 0.9999992 1.1178876637117433E-06 1 +489 1 -1.1626654 0.2381833 2.0698558030857575 0 +492 0 -5.200121 0.005485639 0.0079358908184130224 0 +493 1 10.157462 0.9999612 5.5981438976225112E-05 1 +495 0 -5.5660667 0.003810919 0.0055084966802122675 0 +497 0 -5.8252373 0.0029434143 0.0042527109810820028 0 +501 0 -5.137778 0.005836457 0.0084448957687747774 0 +502 0 -4.818385 0.0080150645 0.011609883223908784 0 +504 0 -6.924154 0.0009827684 0.0014185322093419092 0 +507 0 -5.536234 0.0039258692 0.0056749790015580706 0 +510 0 -6.924154 0.0009827684 0.0014185322093419092 0 +513 0 -5.5660667 0.003810919 0.0055084966802122675 0 +514 1 11.33784 0.9999881 1.7198367596743667E-05 1 +517 0 -6.2979584 0.001836678 0.0026522026287495886 0 +519 1 7.511754 0.99945366 0.00078841187748459848 1 +520 0 -6.8618107 0.0010459214 0.0015097352171979678 0 +521 0 -5.763973 0.0031287964 0.0045209752777798685 0 +522 1 5.377742 0.995403 0.0066473717184157453 1 +523 1 7.6095247 0.99950457 0.0007149370135965138 1 +527 0 -4.449239 0.011552442 0.016763668766800108 0 +528 0 -3.9475894 0.018935692 0.027580387169624309 0 +529 0 -5.200121 0.005485639 0.0079358908184130224 0 +531 0 -4.7229767 0.008810367 0.012766997325712421 0 +532 0 -5.60942 0.003649824 0.0052752157516993657 0 +533 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +534 0 -5.8696694 0.0028158552 0.0040681506882989755 0 +535 0 -4.400241 0.012125549 0.017600393456267439 0 +538 0 -5.137778 0.005836457 0.0084448957687747774 0 +539 0 -4.4058867 0.012058107 0.017501904706086979 0 +540 0 -4.3091917 0.013266058 0.019266958411029617 0 +541 0 -6.1922626 0.0020410197 0.0029475781027268637 0 +544 0 -4.7228336 0.008811617 0.012768816484915386 0 +546 1 11.498754 0.99998987 1.4618599387059818E-05 1 +547 0 -6.6205516 0.0013309219 0.0019213933095301896 0 +548 0 -6.254606 0.001917898 0.0027695986242400887 0 +549 1 5.805538 0.9969982 0.0043372081285626496 1 +557 0 -4.821586 0.007989653 0.011572926966128978 0 +558 0 -5.8696694 0.0028158552 0.0040681506882989755 0 +559 0 -4.771832 0.008393806 0.012160812328843666 0 +560 0 -4.083293 0.016572602 0.024109546243338689 0 +561 0 -4.083293 0.016572602 0.024109546243338689 0 +563 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +565 1 13.257531 0.9999983 2.4937505927785606E-06 1 +566 0 -4.401463 0.012110919 0.017579027701504782 0 +569 1 10.794381 0.9999795 2.9581319217091579E-05 1 +577 0 -6.2356153 0.001954596 0.0028226454560947357 0 +578 0 -6.2356153 0.001954596 0.0028226454560947357 0 +581 1 9.044478 0.999882 0.00017027287211533984 1 +582 1 7.144643 0.99921155 0.0011379419170697886 1 +584 0 -3.801066 0.021858469 0.031884865086942736 0 +586 1 14.420947 0.99999946 7.7392213646475052E-07 1 +590 1 4.6785536 0.9907931 0.013344260348788281 1 +593 0 -5.046649 0.0063897558 0.0092480468882330819 0 +594 1 5.3083115 0.9950741 0.0071241419777647412 1 +600 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +602 0 -5.137778 0.005836457 0.0084448957687747774 0 +604 1 6.8201914 0.99890965 0.0015738975199532537 1 +606 0 -5.506924 0.0040421663 0.0058434312870933374 0 +607 0 -6.924154 0.0009827684 0.0014185322093419092 0 +609 0 -5.460371 0.004233975 0.0061213025677854908 0 +612 1 18.54631 1 -0 1 +613 0 -5.8675113 0.0028219216 0.0040769273975508761 0 +614 0 -5.9320126 0.0026461165 0.0038225989642288226 0 617 0 ? ? ? 0 -618 0 -5.1377897 0.0058363876 0.00844479508184608 0 -619 0 -4.771845 0.008393699 0.012160656505235928 0 -621 0 0.34156227 0.58457 1.2673226591611777 1 -622 0 -2.987461 0.047995564 0.070959798284359485 0 -624 0 -4.514805 0.010827229 0.01570556826182247 0 -627 0 -4.707575 0.008945891 0.012964267461096687 0 -629 0 -5.82631 0.0029402673 0.0042481575122417245 0 -633 1 3.2754412 0.9635766 0.053528713805404145 1 -634 0 -6.192255 0.0020410353 0.0029476006542974429 0 -638 0 -5.82631 0.0029402673 0.0042481575122417245 0 -639 0 -4.8216143 0.007989431 0.011572603255977185 0 -641 0 -5.503734 0.0040550292 0.0058620640547876936 0 -642 0 -5.503734 0.0040550292 0.0058620640547876936 0 -644 0 -6.5582 0.0014164277 0.0020449217526388465 0 -645 0 -5.503734 0.0040550292 0.0058620640547876936 0 -649 0 -5.503734 0.0040550292 0.0058620640547876936 0 -652 0 -4.5978713 0.009972798 0.014459929215864848 0 -653 0 -5.1377897 0.0058363876 0.00844479508184608 0 -654 0 -4.8152137 0.008040319 0.011646612871512004 0 -656 0 -4.771845 0.008393699 0.012160656505235928 0 -657 0 -1.1812477 0.23482794 0.3861438916493285 0 -660 0 -6.235624 0.0019545793 0.0028226212236700603 0 -661 0 -4.449269 0.011552098 0.016763167178581874 0 -665 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -668 1 2.6502724 0.9340278 0.09846261842678565 1 -670 1 6.6141567 0.99866056 0.001933692382575914 1 -678 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -679 0 -6.5582 0.0014164277 0.0020449217526388465 0 -680 1 17.840647 1 -0 1 -681 1 9.976635 0.9999535 6.7074793056570611E-05 1 -682 0 -3.9093504 0.019659286 0.028644855410159493 0 -683 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -685 0 -6.9241447 0.0009827773 0.0014185451543530909 0 -688 0 -5.82631 0.0029402673 0.0042481575122417245 0 -689 0 -3.715599 0.023762457 0.034695860720295138 0 -691 1 4.1132803 0.98390913 0.023403014701784863 1 -692 0 -6.192255 0.0020410353 0.0029476006542974429 0 -693 0 -5.0453706 0.0063978774 0.0092598392764977125 0 -694 0 -5.667263 0.003445403 0.004979248445375453 0 -696 1 6.99251 0.9990821 0.0013248745635289082 1 -697 1 5.7249184 0.996747 0.0047007123976836812 1 -698 1 7.0488224 0.99913234 0.0012523190995148864 1 -0 0 -3.958743 0.018729595 0.027277346289381041 0 -1 0 1.4086351 0.8035506 2.3477703407790238 1 -2 0 -4.911829 0.007305257 0.010577942588486074 0 -3 0 1.5250473 0.82128054 2.4842313552596971 1 -4 0 -4.0157547 0.017710043 0.025779146101163893 0 -7 0 -5.5262413 0.0039651394 0.00573185826626445 0 -12 1 -1.0070925 0.26754925 1.9021236287426191 0 -13 0 -5.864915 0.0028292371 0.0040875113950582112 0 -14 1 6.7670536 0.9988502 0.0016597270056659794 1 -15 1 0.7603264 0.6814246 0.55337402639454369 1 -16 0 -4.9443645 0.0070730555 0.010240520805639591 0 -17 0 -4.595043 0.010000759 0.014500675476667457 0 -19 0 -3.322443 0.034809235 0.051113983778225971 0 -22 0 -5.5806646 0.003755899 0.0054288181764827088 0 +618 0 -5.137778 0.005836457 0.0084448957687747774 0 +619 0 -4.771832 0.008393806 0.012160812328843666 0 +621 0 0.34163857 0.58458847 1.2673868285835368 1 +622 0 -2.9874344 0.047996785 0.07096164998088024 0 +624 0 -4.514783 0.010827464 0.015705910558834784 0 +627 0 -4.707648 0.008945244 0.012963326575909557 0 +629 0 -5.826317 0.0029402478 0.0042481292131313943 0 +633 1 3.275508 0.96357894 0.053525233378840388 1 +634 0 -6.1922626 0.0020410197 0.0029475781027268637 0 +638 0 -5.826317 0.0029402478 0.0042481292131313943 0 +639 0 -4.821586 0.007989653 0.011572926966128978 0 +641 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +642 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +644 0 -6.5582085 0.0014164156 0.0020449042608751558 0 +645 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +649 0 -5.5037236 0.0040550716 0.0058621254381582806 0 +652 0 -4.597859 0.00997292 0.014460107002399077 0 +653 0 -5.137778 0.005836457 0.0084448957687747774 0 +654 0 -4.8151846 0.008040551 0.011646950143324645 0 +656 0 -4.771832 0.008393806 0.012160812328843666 0 +657 0 -1.181243 0.23482879 0.38614549308921986 0 +660 0 -6.2356153 0.001954596 0.0028226454560947357 0 +661 0 -4.449239 0.011552442 0.016763668766800108 0 +665 0 -6.924154 0.0009827684 0.0014185322093419092 0 +668 1 2.650382 0.9340345 0.098452215112417155 1 +670 1 6.61419 0.9986606 0.0019336062759187284 1 +678 0 -6.924154 0.0009827684 0.0014185322093419092 0 +679 0 -6.5582085 0.0014164156 0.0020449042608751558 0 +680 1 17.840736 1 -0 1 +681 1 9.976583 0.9999535 6.7074793056570611E-05 1 +682 0 -3.9093199 0.019659873 0.028645718862376924 0 +683 0 -6.924154 0.0009827684 0.0014185322093419092 0 +685 0 -6.924154 0.0009827684 0.0014185322093419092 0 +688 0 -5.826317 0.0029402478 0.0042481292131313943 0 +689 0 -3.715537 0.023763895 0.034697985758682486 0 +691 1 4.1131887 0.98390764 0.023405199644172255 1 +692 0 -6.1922626 0.0020410197 0.0029475781027268637 0 +693 0 -5.04535 0.0063980077 0.0092600285937630804 0 +694 0 -5.6672792 0.0034453473 0.0049791678868571602 0 +696 1 6.992524 0.99908215 0.0013247884932010956 1 +697 1 5.7250013 0.99674726 0.0047003673098582315 1 +698 1 7.048875 0.9991324 0.0012522330335155807 1 +0 0 -3.9587312 0.018729815 0.027277669434812288 0 +1 0 1.4087048 0.80356157 2.3478508849071225 1 +2 0 -4.9118805 0.0073048836 0.010577399834229673 0 +3 0 1.5250807 0.8212854 2.4842708103087876 1 +4 0 -4.015728 0.017710507 0.025779827285120106 0 +7 0 -5.5262666 0.0039650393 0.0057317132527183491 0 +12 1 -1.007082 0.2675513 1.9021125403565939 0 +13 0 -5.8650293 0.0028289144 0.0040870445117997438 0 +14 1 6.767168 0.99885035 0.0016595548250566607 1 +15 1 0.7605486 0.68147284 0.5532719395031015 1 +16 0 -4.9443603 0.0070730853 0.010240564107579979 0 +17 0 -4.5950603 0.010000589 0.014500428468588475 0 +19 0 -3.322402 0.034810614 0.051116044045370665 0 +22 0 -5.5806894 0.0037558062 0.005428683982832385 0 23 1 ? ? ? 0 -24 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -26 0 -5.5484257 0.003878481 0.0056063443785696271 0 -27 0 -4.3080635 0.013280834 0.019288562964927369 0 -29 0 -6.5197806 0.001471823 0.0021249558559103803 0 -30 0 -5.5806646 0.003755899 0.0054288181764827088 0 -33 0 -5.577999 0.003765886 0.0054432808185844294 0 -34 0 -5.2392616 0.005276186 0.0076320791567411374 0 -36 1 8.057589 0.99968344 0.00045677223115128229 1 -38 1 4.1148243 0.9839335 0.023367269514222317 1 -39 1 -0.31325245 0.42232105 1.2435879322034293 0 -42 1 5.3922577 0.9954689 0.0065518292521607709 1 -43 1 -0.983716 0.27215508 1.8774991492725446 0 -47 0 -6.853265 0.0010548885 0.0015226856442652843 0 -49 1 3.7823315 0.97773737 0.032481104136983185 1 -53 1 4.7163496 0.99113154 0.012851548630617684 1 -55 1 4.0095263 0.98218125 0.025938811833612716 1 -57 1 0.88929653 0.708745 0.49666143702034871 1 -58 1 0.330575 0.5818993 0.78115862070227093 1 -59 1 1.3888245 0.8004045 0.32119883631671436 1 -61 0 -6.5337505 0.0014514343 0.0020954982240835363 0 -62 1 4.610319 0.9901494 0.014281901247611249 1 -65 1 2.396164 0.9165343 0.12573921501988825 1 -67 1 2.0528374 0.886234 0.17424044233274416 1 -75 0 -5.2814636 0.0050592567 0.0073174907083955508 0 -78 0 -4.592314 0.010027818 0.014540109094429828 0 -80 0 -4.096719 0.016355203 0.023790656143087566 0 -81 0 -4.6547194 0.009426872 0.013664610783443085 0 -83 0 -3.5145454 0.02890119 0.042309995874323041 0 -84 1 5.9317665 0.99735326 0.003823506963331196 1 -85 1 2.8984337 0.9477689 0.07739273307849516 1 -86 1 0.63173294 0.6528823 0.61510521384483363 1 -87 1 3.8592606 0.97935176 0.030100961560428498 1 -89 0 -5.9273205 0.0026585283 0.0038405529338516237 0 -94 0 -5.8676443 0.0028215474 0.0040763860729673583 0 -101 1 -2.1323166 0.10599527 3.2379282406619296 0 -103 1 0.7015152 0.6685236 0.58094958332490176 1 -107 1 3.5015984 0.9707332 0.042853222879590994 1 -110 0 -4.5247397 0.010721342 0.015551140826768391 0 -114 0 -4.186434 0.014972803 0.021764536641712809 0 -116 0 -0.7708025 0.31630555 0.54857637650500712 0 -118 0 -6.5635567 0.0014088711 0.0020340045786757435 0 -119 0 -4.87959 0.007542803 0.010923212217175612 0 -124 1 5.1567287 0.9942725 0.0082868327018022318 1 -126 1 5.7151337 0.9967151 0.0047468686381897125 1 -127 0 -5.2313433 0.005317908 0.0076925916064113616 0 -130 0 -3.6252584 0.025950823 0.037933483100260706 0 -134 0 -5.5806646 0.003755899 0.0054288181764827088 0 -135 0 -3.2147431 0.038614668 0.056813302586015808 0 -136 0 -4.9443645 0.0070730555 0.010240520805639591 0 +24 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +26 0 -5.5484886 0.003878238 0.005605992329824022 0 +27 0 -4.308031 0.013281259 0.019289183899785744 0 +29 0 -6.5197434 0.0014718776 0.0021250347414157422 0 +30 0 -5.5806894 0.0037558062 0.005428683982832385 0 +33 0 -5.5780525 0.0037656857 0.0054429908495141413 0 +34 0 -5.2392373 0.0052763135 0.0076322642083003767 0 +36 1 8.057798 0.9996835 0.00045668621259835828 1 +38 1 4.114894 0.98393464 0.023365609001363471 1 +39 1 -0.3135004 0.42226055 1.2437946172532774 0 +42 1 5.392254 0.9954689 0.0065518292521607709 1 +43 1 -0.9835901 0.27218 1.877367082182928 0 +47 0 -6.8533473 0.0010548015 0.0015225600517844947 0 +49 1 3.7823277 0.9777373 0.032481192086294464 1 +53 1 4.716385 0.9911319 0.012851028066155212 1 +55 1 4.0097027 0.98218435 0.025934259168957082 1 +57 1 0.88941383 0.7087692 0.4966121782855778 1 +58 1 0.33055496 0.5818944 0.78117073846642104 1 +59 1 1.3889704 0.8004278 0.32115682990718269 1 +61 0 -6.5338383 0.0014513072 0.0020953145541365717 0 +62 1 4.610484 0.990151 0.01427955638536158 1 +65 1 2.396165 0.91653436 0.12573912119762518 1 +67 1 2.0527573 0.88622594 0.17425354144858277 1 +75 0 -5.281555 0.0050587957 0.0073168222374257793 0 +78 0 -4.5923715 0.010027246 0.01453927440162823 0 +80 0 -4.0968666 0.016352825 0.023787167498412636 0 +81 0 -4.654694 0.009427108 0.013664953952957026 0 +83 0 -3.5146813 0.028897377 0.042304331417997096 0 +84 1 5.931757 0.9973532 0.0038235931828602401 1 +85 1 2.8981676 0.94775575 0.077412784606301688 1 +86 1 0.63182735 0.6529037 0.61505793061937375 1 +87 1 3.8591766 0.97935003 0.030103507888257507 1 +89 0 -5.927352 0.002658445 0.0038404323598130001 0 +94 0 -5.8677187 0.0028213381 0.0040760832411949836 0 +101 1 -2.1323013 0.10599671 3.237908567373958 0 +103 1 0.70157814 0.66853756 0.58091948452216657 1 +107 1 3.5016832 0.9707356 0.042849679528335009 1 +110 0 -4.5249724 0.010718874 0.015547541665048462 0 +114 0 -4.1865177 0.014971565 0.021762723836440984 0 +116 0 -0.77069855 0.31632802 0.54862379426941132 0 +118 0 -6.5636287 0.0014087697 0.0020338580862699291 0 +119 0 -4.8796797 0.007542132 0.01092223678579787 0 +124 1 5.156763 0.99427265 0.0082865732417842213 1 +126 1 5.7150803 0.99671495 0.0047471274623934172 1 +127 0 -5.2313895 0.005317663 0.0076922363466130801 0 +130 0 -3.6251273 0.025954138 0.037938393812913335 0 +134 0 -5.5806894 0.0037558062 0.005428683982832385 0 +135 0 -3.2147598 0.038614046 0.056812369001802629 0 +136 0 -4.9443603 0.0070730853 0.010240564107579979 0 139 0 ? ? ? 0 -140 0 -5.929985 0.0026514728 0.0038303469497138333 0 -142 1 3.366704 0.96664757 0.048938108369594384 1 -143 0 -5.924527 0.002665945 0.0038512817060373356 0 -146 1 -0.8934479 0.2903988 1.7838925716508713 0 -148 0 -2.4677153 0.07815268 0.11740026759678464 0 -149 1 7.946436 0.9996462 0.00051053483006131134 1 -153 0 -4.6546555 0.009427468 0.013665478880395175 0 -155 1 1.8596773 0.86525935 0.20879546972753268 1 -157 0 -5.8676443 0.0028215474 0.0040763860729673583 0 +140 0 -5.929989 0.0026514626 0.0038303321306625395 0 +142 1 3.3668652 0.9666528 0.048930280060313358 1 +143 0 -5.92461 0.0026657244 0.0038509624182418525 0 +146 1 -0.8934498 0.29039842 1.783894496397034 0 +148 0 -2.4678388 0.07814378 0.11738634539611086 0 +149 1 7.9463863 0.9996462 0.00051053483006131134 1 +153 0 -4.6546416 0.009427598 0.013665667420271003 0 +155 1 1.8597918 0.8652727 0.20877320830001908 1 +157 0 -5.8677187 0.0028213381 0.0040760832411949836 0 158 0 ? ? ? 0 -159 1 9.591583 0.9999317 9.8549424786689596E-05 1 -160 1 7.3637104 0.9993666 0.00091411935601581808 1 -162 0 -5.2313433 0.005317908 0.0076925916064113616 0 -163 0 -5.179269 0.0056005795 0.0081026393177347153 0 -165 0 -4.0632863 0.016901843 0.024592626249255688 0 -166 1 5.4290752 0.995632 0.0063155054515529937 1 -168 0 -5.2313433 0.005317908 0.0076925916064113616 0 -170 0 -5.929985 0.0026514728 0.0038303469497138333 0 -172 0 -6.853265 0.0010548885 0.0015226856442652843 0 -175 1 4.900942 0.99261534 0.010693341382165791 1 -178 0 -4.595043 0.010000759 0.014500675476667457 0 -182 0 -3.322443 0.034809235 0.051113983778225971 0 -184 1 4.293812 0.9865311 0.019563593815186388 1 -185 0 -5.8676443 0.0028215474 0.0040763860729673583 0 -186 1 3.007121 0.9528948 0.069611135731289506 1 -190 1 10.157112 0.9999612 5.5981438976225112E-05 1 -193 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -194 0 -5.2313433 0.005317908 0.0076925916064113616 0 -195 0 -4.595043 0.010000759 0.014500675476667457 0 -197 0 -3.0898094 0.04352957 0.064207727779592202 0 -200 1 8.11038 0.9996997 0.00043328935662039213 1 -203 0 -3.958743 0.018729595 0.027277346289381041 0 -208 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -213 1 11.922752 0.9999934 9.5450686985595153E-06 1 -214 1 11.574423 0.9999906 1.3586693394834943E-05 1 -215 1 6.3406 0.9982399 0.0025415612296557318 1 -217 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -220 0 -6.2143 0.0019966215 0.002883395404928794 0 -221 1 7.86919 0.9996178 0.00055148176940528562 1 -222 1 -3.087016 0.043646015 4.518006244394094 0 -224 1 8.318844 0.99975616 0.00035183341007045081 1 -225 0 -6.853265 0.0010548885 0.0015226856442652843 0 -227 1 6.046664 0.99763983 0.0034090230260584603 1 -229 1 10.046562 0.99995667 6.2517048110806339E-05 1 -230 1 4.7035074 0.991018 0.01301683734542588 1 -231 1 5.973811 0.9974619 0.0036663373280547808 1 -232 0 1.1069822 0.75156605 2.009065769015387 1 -234 0 -3.7642503 0.022659624 0.033067001215731306 0 +159 1 9.591633 0.9999317 9.8549424786689596E-05 1 +160 1 7.3638344 0.99936664 0.00091403331018992775 1 +162 0 -5.2313895 0.005317663 0.0076922363466130801 0 +163 0 -5.179346 0.0056001497 0.0081020157474404574 0 +165 0 -4.06332 0.01690128 0.02459180075396205 0 +166 1 5.429183 0.99563247 0.0063148145030462359 1 +168 0 -5.2313895 0.005317663 0.0076922363466130801 0 +170 0 -5.929989 0.0026514626 0.0038303321306625395 0 +172 0 -6.8533473 0.0010548015 0.0015225600517844947 0 +175 1 4.9009514 0.99261546 0.010693168120043763 1 +178 0 -4.5950603 0.010000589 0.014500428468588475 0 +182 0 -3.322402 0.034810614 0.051116044045370665 0 +184 1 4.2938995 0.9865323 0.019561850509266822 1 +185 0 -5.8677187 0.0028213381 0.0040760832411949836 0 +186 1 3.0070934 0.95289356 0.06961303081877146 1 +190 1 10.157295 0.9999612 5.5981438976225112E-05 1 +193 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +194 0 -5.2313895 0.005317663 0.0076922363466130801 0 +195 0 -4.5950603 0.010000589 0.014500428468588475 0 +197 0 -3.0897326 0.043532766 0.064212548934943228 0 +200 1 8.110437 0.9996997 0.00043328935662039213 1 +203 0 -3.9587312 0.018729815 0.027277669434812288 0 +208 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +213 1 11.922953 0.9999934 9.5450686985595153E-06 1 +214 1 11.574621 0.9999906 1.3586693394834943E-05 1 +215 1 6.3407736 0.9982402 0.0025411305149810205 1 +217 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +220 0 -6.214381 0.00199646 0.0028831618214611074 0 +221 1 7.869068 0.99961776 0.00055156779361044005 1 +222 1 -3.0869913 0.04364705 4.5179720125966218 0 +224 1 8.318922 0.9997562 0.00035174739777412068 1 +225 0 -6.8533473 0.0010548015 0.0015225600517844947 0 +227 1 6.0467644 0.9976401 0.0034086782470624751 1 +229 1 10.046583 0.99995667 6.2517048110806339E-05 1 +230 1 4.7036476 0.99101925 0.01301501516187459 1 +231 1 5.973727 0.99746174 0.0036665959584801975 1 +232 0 1.1072111 0.7516088 2.0093139681229384 1 +234 0 -3.7643795 0.022656763 0.033062777940319171 0 235 0 ? ? ? 0 -236 1 8.480511 0.9997926 0.00029928085261473974 1 -238 1 9.448708 0.9999212 0.00011368501131740873 1 -243 0 -4.5272436 0.010694817 0.015512459136733886 0 -245 0 -3.3495836 0.033908803 0.04976871201005937 0 -251 1 7.314123 0.9993344 0.0009605848516411478 1 -253 1 5.3922577 0.9954689 0.0065518292521607709 1 -255 1 2.135953 0.8943488 0.1610904959698177 1 -256 0 -5.929985 0.0026514728 0.0038303469497138333 0 -261 1 7.9641056 0.9996524 0.00050158859464882689 1 -263 1 7.373728 0.99937284 0.0009050845723178523 1 -264 1 3.329814 0.9654376 0.050745093044825204 1 -265 0 -2.7222924 0.061670676 0.091833742616940522 0 -266 1 5.63109 0.99642813 0.0051623412667997489 1 -270 1 4.851303 0.99224246 0.011235407155212151 1 -273 1 -0.24666595 0.4386443 1.1888766064173391 0 -274 0 -5.019101 0.0065670544 0.0095055027075708533 0 -281 0 -5.577999 0.003765886 0.0054432808185844294 0 -282 1 1.5178642 0.82022375 0.2859105789563019 1 -286 1 11.969433 0.9999937 9.115109290810302E-06 1 -287 0 -5.5806646 0.003755899 0.0054288181764827088 0 -289 1 5.6194344 0.9963864 0.005222752234881816 1 +236 1 8.480574 0.9997926 0.00029928085261473974 1 +238 1 9.448872 0.9999212 0.00011368501131740873 1 +243 0 -4.5274115 0.010693042 0.015509870525144425 0 +245 0 -3.349556 0.03390971 0.049770063842961822 0 +251 1 7.3143826 0.9993346 0.00096032670586479658 1 +253 1 5.392254 0.9954689 0.0065518292521607709 1 +255 1 2.1359367 0.8943473 0.161092899713107 1 +256 0 -5.929989 0.0026514626 0.0038303321306625395 0 +261 1 7.964054 0.9996524 0.00050158859464882689 1 +263 1 7.373974 0.999373 0.00090482643647214148 1 +264 1 3.329836 0.9654383 0.050744024207717064 1 +265 0 -2.7221441 0.06167926 0.09184693927101252 0 +266 1 5.631093 0.99642813 0.0051623412667997489 1 +270 1 4.8514376 0.9922435 0.011233847210853399 1 +273 1 -0.24652672 0.43867856 1.1887638884777549 0 +274 0 -5.0191584 0.0065666814 0.0095049610328734775 0 +281 0 -5.5780525 0.0037656857 0.0054429908495141413 0 +282 1 1.5177917 0.8202131 0.28592934523487895 1 +286 1 11.969671 0.9999937 9.115109290810302E-06 1 +287 0 -5.5806894 0.0037558062 0.005428683982832385 0 +289 1 5.619467 0.9963865 0.0052225796285123843 1 292 1 ? ? ? 0 294 0 ? ? ? 0 -295 1 5.1557827 0.9942671 0.008294616524040609 1 -298 0 -2.7162447 0.062021572 0.092373351657769942 0 -302 1 12.148464 0.9999947 7.6532482629398447E-06 1 -305 1 6.6747093 0.9987391 0.0018202082685338874 1 -306 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -307 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -310 0 -6.2169647 0.001991319 0.0028757302519144814 0 -313 0 -7.202586 0.0007441034 0.0010739138801475331 0 +295 1 5.15592 0.9942679 0.0082934921915667707 1 +298 0 -2.7161427 0.062027507 0.092382479307369897 0 +302 1 12.148573 0.9999947 7.6532482629398447E-06 1 +305 1 6.6746464 0.99873906 0.0018202943684232252 1 +306 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +307 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +310 0 -6.217018 0.0019912128 0.002875576774251216 0 +313 0 -7.202647 0.000744058 0.0010738483301681407 0 315 0 ? ? ? 0 -318 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -320 1 4.269348 0.9862021 0.020044739597648867 1 -322 0 -5.2313433 0.005317908 0.0076925916064113616 0 -324 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -325 0 -4.782538 0.008305164 0.012031851693160102 0 -326 1 3.4337597 0.9687431 0.045813986417611863 1 -330 1 4.9421005 0.99291104 0.010263628664885187 1 -334 1 4.433858 0.9882706 0.017022000356151753 1 -335 0 -7.202586 0.0007441034 0.0010739138801475331 0 -337 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -339 1 4.252741 0.9859743 0.020378035180971199 1 -340 1 5.3041697 0.99505377 0.0071536104782971939 1 -341 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -342 0 -6.5662856 0.001405037 0.0020284652923748298 0 -345 0 -7.202586 0.0007441034 0.0010739138801475331 0 -351 0 -5.8676443 0.0028215474 0.0040763860729673583 0 -356 1 -1.7830639 0.14392522 2.7966086796315368 0 -357 1 9.025284 0.99987966 0.00017362693354248715 1 -359 1 4.6527195 0.99055445 0.013691810950701988 1 -362 0 -4.567993 0.010272155 0.014896226309139005 0 -363 0 -2.8947659 0.052412912 0.077669554989392792 0 -364 0 -5.8676443 0.0028215474 0.0040763860729673583 0 -365 0 -6.2169647 0.001991319 0.0028757302519144814 0 -367 1 8.160849 0.99971443 0.00041204327568713356 1 -369 0 -6.1704597 0.0020859146 0.0030124816358925734 0 -372 0 -4.6029615 0.009922665 0.014386875733067683 0 -374 0 -5.2392616 0.005276186 0.0076320791567411374 0 -375 0 -7.202586 0.0007441034 0.0010739138801475331 0 -380 0 -7.202586 0.0007441034 0.0010739138801475331 0 -382 0 -4.3558154 0.012669398 0.018394851790680372 0 -385 0 -4.6573367 0.009402462 0.013629059949172623 0 -386 1 4.0342703 0.9826092 0.025310329827161745 1 -390 0 -6.5636206 0.0014087812 0.0020338745687672002 0 -393 0 -7.202586 0.0007441034 0.0010739138801475331 0 -394 0 -6.138221 0.002154111 0.0031110771644373997 0 -397 0 -5.293685 0.0049981093 0.007228827828065121 0 -400 1 5.540103 0.9960892 0.0056531241255807999 1 -401 0 -5.929985 0.0026514728 0.0038303469497138333 0 -402 0 -3.5607805 0.027631443 0.040424852471735717 0 -403 0 -4.9714413 0.0068854108 0.0099679040201330182 0 -405 0 -6.853265 0.0010548885 0.0015226856442652843 0 -407 0 -6.853265 0.0010548885 0.0015226856442652843 0 -408 0 -4.852604 0.0077475267 0.011220841646650362 0 -410 0 -6.853265 0.0010548885 0.0015226856442652843 0 +318 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +320 1 4.2691555 0.9861995 0.020048576157386616 1 +322 0 -5.2313895 0.005317663 0.0076922363466130801 0 +324 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +325 0 -4.7825737 0.00830487 0.012031423555298607 0 +326 1 3.4337902 0.96874404 0.045812566164347868 1 +330 1 4.942357 0.9929128 0.010261030509225668 1 +334 1 4.4340143 0.9882724 0.017019302989005319 1 +335 0 -7.202647 0.000744058 0.0010738483301681407 0 +337 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +339 1 4.2528763 0.9859762 0.020375244317437313 1 +340 1 5.3041573 0.9950537 0.0071536968970724989 1 +341 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +342 0 -6.566318 0.0014049915 0.002028399530822185 0 +345 0 -7.202647 0.000744058 0.0010738483301681407 0 +351 0 -5.8677187 0.0028213381 0.0040760832411949836 0 +356 1 -1.7830734 0.14392404 2.7966204797568066 0 +357 1 9.025423 0.9998797 0.00017354093187002464 1 +359 1 4.652974 0.99055684 0.01368833850264925 1 +362 0 -4.568018 0.010271903 0.014895859768114864 0 +363 0 -2.8947177 0.052415304 0.077673196244462464 0 +364 0 -5.8677187 0.0028213381 0.0040760832411949836 0 +365 0 -6.217018 0.0019912128 0.002875576774251216 0 +367 1 8.160868 0.9997145 0.00041195725980106571 1 +369 0 -6.1704435 0.0020859484 0.003012530443726541 0 +372 0 -4.602908 0.00992319 0.014387641126567057 0 +374 0 -5.2392373 0.0052763135 0.0076322642083003767 0 +375 0 -7.202647 0.000744058 0.0010738483301681407 0 +380 0 -7.202647 0.000744058 0.0010738483301681407 0 +382 0 -4.3558393 0.0126691 0.018394416316927616 0 +385 0 -4.657366 0.009402191 0.013628665246238838 0 +386 1 4.0343037 0.98260975 0.025309542208128807 1 +390 0 -6.563681 0.001408696 0.0020337516227971952 0 +393 0 -7.202647 0.000744058 0.0010738483301681407 0 +394 0 -6.1382427 0.0021540637 0.0031110088288026387 0 +397 0 -5.29366 0.0049982327 0.0072290067512698264 0 +400 1 5.5399637 0.9960887 0.0056539010862402504 1 +401 0 -5.929989 0.0026514626 0.0038303321306625395 0 +402 0 -3.5607262 0.027632903 0.040427019128647733 0 +403 0 -4.971462 0.0068852706 0.0099677004041916865 0 +405 0 -6.8533473 0.0010548015 0.0015225600517844947 0 +407 0 -6.8533473 0.0010548015 0.0015225600517844947 0 +408 0 -4.8526897 0.007746867 0.011219882263273698 0 +410 0 -6.8533473 0.0010548015 0.0015225600517844947 0 411 0 ? ? ? 0 -412 1 7.5878735 0.9994937 0.0007305952773892084 1 -417 0 -6.853265 0.0010548885 0.0015226856442652843 0 -420 0 -3.6393833 0.025596164 0.037408282326143294 0 -421 1 9.157879 0.9998946 0.00015204067460350784 1 -424 0 -5.929985 0.0026514728 0.0038303469497138333 0 -425 1 11.787092 0.9999924 1.1006931643385188E-05 1 -426 0 -2.8858938 0.0528553 0.078343247179589737 0 -427 1 3.7513933 0.9770539 0.033489971523710152 1 -431 0 -3.1672115 0.040418427 0.059522641953831591 0 -432 0 -4.0958214 0.016369646 0.023811839527671772 0 -433 0 -4.9770546 0.006847133 0.0099122989947988335 0 -435 1 6.0953283 0.9977517 0.0032472445338206329 1 -437 0 -5.293685 0.0049981093 0.007228827828065121 0 -438 0 -4.0315905 0.017436652 0.025377670109658607 0 -443 0 -6.883071 0.001023942 0.0014779928968589077 0 -444 0 -2.8598413 0.05417483 0.080354558930328251 0 -445 0 -6.5662856 0.001405037 0.0020284652923748298 0 -446 0 -7.202586 0.0007441034 0.0010739138801475331 0 -447 0 -4.657385 0.009402014 0.013628407536477802 0 -448 0 -7.202586 0.0007441034 0.0010739138801475331 0 -458 0 -4.315982 0.0131774675 0.019137437274964097 0 -459 0 -3.9745798 0.018440744 0.026852730008789516 0 -460 0 -4.0780954 0.016657524 0.024234132391009727 0 -461 0 -4.651926 0.009452992 0.013702652909678398 0 -462 0 -3.4417953 0.031014483 0.045452993041962333 0 -463 0 -5.0814424 0.0061726063 0.0089327861708581625 0 -468 0 -5.293685 0.0049981093 0.007228827828065121 0 -469 0 -6.5118628 0.0014835056 0.0021418352657478924 0 -470 0 -5.5806646 0.003755899 0.0054288181764827088 0 -471 0 -3.4417953 0.031014483 0.045452993041962333 0 -472 0 -4.021085 0.017617555 0.025643315145556417 0 -473 0 -5.293685 0.0049981093 0.007228827828065121 0 -475 0 -5.929985 0.0026514728 0.0038303469497138333 0 -476 0 -4.952282 0.007017668 0.010160046395133819 0 -477 0 -5.293685 0.0049981093 0.007228827828065121 0 -478 0 -4.625146 0.009707074 0.014072760766848626 0 -479 1 5.407731 0.9955382 0.0064514560081750582 1 -481 0 -3.3825712 0.032844618 0.048180404889341902 0 -485 0 -5.984267 0.0025117334 0.0036282238597670231 0 -486 0 -5.5806646 0.003755899 0.0054288181764827088 0 -488 1 -0.16338348 0.45924476 1.1226648412878133 0 -490 0 -7.202586 0.0007441034 0.0010739138801475331 0 -491 1 4.1399994 0.9843267 0.02279083681171163 1 -494 0 -0.08180237 0.4795608 0.9421984425687574 0 -496 0 -7.202586 0.0007441034 0.0010739138801475331 0 -498 0 -4.9443645 0.0070730555 0.010240520805639591 0 -499 0 -4.9443645 0.0070730555 0.010240520805639591 0 -500 0 -3.322443 0.034809235 0.051113983778225971 0 -503 0 -4.595043 0.010000759 0.014500675476667457 0 -505 0 -5.8655076 0.0028275654 0.0040850927662499327 0 -506 1 8.127453 0.9997048 0.00042597791694096366 1 -508 0 -4.657385 0.009402014 0.013628407536477802 0 -509 0 -6.5662856 0.001405037 0.0020284652923748298 0 -511 0 -4.3080635 0.013280834 0.019288562964927369 0 -512 0 -4.657385 0.009402014 0.013628407536477802 0 -515 1 5.3389215 0.9952219 0.0069098437158003059 1 -516 0 -7.202586 0.0007441034 0.0010739138801475331 0 -518 0 -5.2989383 0.004972052 0.0071910464961066779 0 -524 0 -5.5806646 0.003755899 0.0054288181764827088 0 -525 0 -5.640341 0.0035390882 0.0051148813258727784 0 -526 0 -5.293685 0.0049981093 0.007228827828065121 0 -530 1 4.635607 0.990393 0.013927001947701955 1 -536 0 -3.958743 0.018729595 0.027277346289381041 0 -537 0 -3.61734 0.026151732 0.038231085920625615 0 -542 0 -4.592907 0.0100219315 0.014531530104124092 0 -543 0 -4.9443645 0.0070730555 0.010240520805639591 0 -545 0 -4.3080635 0.013280834 0.019288562964927369 0 -550 0 -5.5806646 0.003755899 0.0054288181764827088 0 -551 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -552 0 -4.0064945 0.017871857 0.026016822223623635 0 -553 0 -3.0457454 0.045401514 0.067034045980015888 0 -554 0 -5.929985 0.0026514728 0.0038303469497138333 0 -555 0 -2.695547 0.063236624 0.094243422015334713 0 -556 0 -3.6447773 0.025461977 0.037209620002719525 0 -562 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -564 0 -4.382801 0.01233624 0.017908120445798863 0 -567 0 -4.305335 0.013316637 0.019340911707696886 0 -568 1 2.6288052 0.9326926 0.10052644333787665 1 -570 1 4.967929 0.9930905 0.010002883771278568 1 -571 1 8.770433 0.9998448 0.00022393879069825088 1 -572 0 -5.5806646 0.003755899 0.0054288181764827088 0 -573 0 -6.853265 0.0010548885 0.0015226856442652843 0 -574 1 5.098628 0.99393195 0.0087810157366521747 1 -575 0 -3.61734 0.026151732 0.038231085920625615 0 -576 0 -4.3080635 0.013280834 0.019288562964927369 0 -579 0 -6.5039444 0.0014952815 0.0021588496039274389 0 -580 0 -3.9666605 0.018584637 0.027064239426035003 0 -583 0 -5.929985 0.0026514728 0.0038303469497138333 0 -585 0 -7.202586 0.0007441034 0.0010739138801475331 0 -587 0 -4.0958214 0.016369646 0.023811839527671772 0 -588 1 4.507454 0.9890938 0.015820778884382193 1 -589 0 -4.657385 0.009402014 0.013628407536477802 0 -591 1 3.0920782 0.9565648 0.064065415429797629 1 -592 1 4.694976 0.9909417 0.013127908117020385 1 -595 0 -4.3080635 0.013280834 0.019288562964927369 0 -596 0 -4.6029615 0.009922665 0.014386875733067683 0 -597 0 -3.275938 0.036405943 0.053502599655809731 0 -598 0 -5.5806646 0.003755899 0.0054288181764827088 0 -599 0 -3.6009378 0.026572725 0.038854895740116642 0 -601 0 -6.853265 0.0010548885 0.0015226856442652843 0 -603 1 3.5292854 0.97150964 0.041699790332770832 1 -605 1 8.488092 0.9997941 0.00029704461603679162 1 -608 1 7.331711 0.99934596 0.00094389151985603862 1 -610 1 7.064438 0.99914575 0.001232954379056525 1 -611 1 4.9259205 0.99279624 0.01043044005182185 1 -615 0 -4.2615585 0.013904255 0.020200362788656924 0 -616 0 -5.5806646 0.003755899 0.0054288181764827088 0 -620 0 -5.5806646 0.003755899 0.0054288181764827088 0 -623 0 -7.202586 0.0007441034 0.0010739138801475331 0 -625 0 -4.288636 0.01353784 0.019664383632128662 0 -626 1 3.0649357 0.955423 0.065788491321487508 1 -628 0 -6.5662856 0.001405037 0.0020284652923748298 0 -630 0 -3.3060408 0.035364535 0.051944242424523938 0 -631 0 -4.3080635 0.013280834 0.019288562964927369 0 -632 0 -7.202586 0.0007441034 0.0010739138801475331 0 -635 0 -4.905778 0.0073492695 0.010641907810662122 0 -636 1 8.309145 0.99975383 0.00035518789362715982 1 -637 0 -2.9891138 0.047920104 0.070845449338562141 0 -640 0 -4.7143965 0.008885614 0.012876523744768983 0 -643 0 -7.202586 0.0007441034 0.0010739138801475331 0 -646 0 -6.828944 0.0010808305 0.0015601520338859064 0 -647 0 -6.6232967 0.0013272782 0.0019161295898977376 0 -648 1 8.32774 0.99975836 0.0003486509585214664 1 -650 0 -4.681797 0.00917735 0.013301246856517835 0 -651 0 -6.2417374 0.0019426894 0.0028054343107074258 0 -655 0 -5.5806646 0.003755899 0.0054288181764827088 0 -658 1 6.517256 0.9985245 0.002130287281102739 1 -659 0 -7.202586 0.0007441034 0.0010739138801475331 0 -662 0 -6.1704597 0.0020859146 0.0030124816358925734 0 -663 0 -6.1704597 0.0020859146 0.0030124816358925734 0 -664 0 -5.001375 0.0066837147 0.0096749304095307062 0 -666 0 -3.811658 0.021633148 0.031552570360073216 0 -667 0 -5.2313433 0.005317908 0.0076925916064113616 0 -669 1 6.017066 0.99756914 0.0035112536325718441 1 -671 0 -4.90946 0.007322456 0.010602938605811105 0 -672 0 -5.8676443 0.0028215474 0.0040763860729673583 0 -673 0 -4.251504 0.014042789 0.020403057803134814 0 -674 0 -6.853265 0.0010548885 0.0015226856442652843 0 -675 0 -4.269477 0.013796104 0.020042142841059332 0 -676 0 -6.5118628 0.0014835056 0.0021418352657478924 0 -677 0 -4.657385 0.009402014 0.013628407536477802 0 -684 0 -7.202586 0.0007441034 0.0010739138801475331 0 -686 0 -7.202586 0.0007441034 0.0010739138801475331 0 -687 0 -5.1561794 0.0057306485 0.0082913583947259867 0 -690 0 -6.6232967 0.0013272782 0.0019161295898977376 0 -695 0 -6.5662856 0.001405037 0.0020284652923748298 0 +412 1 7.58807 0.9994938 0.00073050924250844221 1 +417 0 -6.8533473 0.0010548015 0.0015225600517844947 0 +420 0 -3.6393862 0.025596093 0.037408177529044309 0 +421 1 9.158096 0.9998946 0.00015204067460350784 1 +424 0 -5.929989 0.0026514626 0.0038303321306625395 0 +425 1 11.787216 0.9999924 1.1006931643385188E-05 1 +426 0 -2.8856697 0.052866522 0.078360338509909269 0 +427 1 3.7514791 0.9770558 0.033487155179819289 1 +431 0 -3.1671677 0.04042013 0.059525201537593574 0 +432 0 -4.0958 0.016369991 0.023812344938505647 0 +433 0 -4.977004 0.0068474766 0.0099127982067911749 0 +435 1 6.0955677 0.99775225 0.0032464688681806695 1 +437 0 -5.29366 0.0049982327 0.0072290067512698264 0 +438 0 -4.0314236 0.01743951 0.025381868212915427 0 +443 0 -6.8831377 0.0010238737 0.0014778942080997193 0 +444 0 -2.8597002 0.054182064 0.080365593989833936 0 +445 0 -6.566318 0.0014049915 0.002028399530822185 0 +446 0 -7.202647 0.000744058 0.0010738483301681407 0 +447 0 -4.6573305 0.00940252 0.01362914404396972 0 +448 0 -7.202647 0.000744058 0.0010738483301681407 0 +458 0 -4.315879 0.013178807 0.019139395194286708 0 +459 0 -3.9744272 0.018443506 0.026856790044972514 0 +460 0 -4.077998 0.016659118 0.024236471626704163 0 +461 0 -4.6519527 0.009452742 0.013702288028212783 0 +462 0 -3.4416695 0.031018266 0.04545862550240317 0 +463 0 -5.0814285 0.0061726915 0.008932909875164487 0 +468 0 -5.29366 0.0049982327 0.0072290067512698264 0 +469 0 -6.511895 0.0014834575 0.0021417657985981662 0 +470 0 -5.5806894 0.0037558062 0.005428683982832385 0 +471 0 -3.4416695 0.031018266 0.04545862550240317 0 +472 0 -4.021002 0.017618991 0.025645424156179257 0 +473 0 -5.29366 0.0049982327 0.0072290067512698264 0 +475 0 -5.929989 0.0026514626 0.0038303321306625395 0 +476 0 -4.9522076 0.007018186 0.010160799401121795 0 +477 0 -5.29366 0.0049982327 0.0072290067512698264 0 +478 0 -4.6251297 0.009707229 0.014072987349939807 0 +479 1 5.4077578 0.9955383 0.0064512832547389109 1 +481 0 -3.3824787 0.032847557 0.048184789348920593 0 +485 0 -5.984297 0.0025116594 0.0036281167734491422 0 +486 0 -5.5806894 0.0037558062 0.005428683982832385 0 +488 1 -0.16332245 0.4592599 1.1226172818192048 0 +490 0 -7.202647 0.000744058 0.0010738483301681407 0 +491 1 4.140068 0.98432773 0.022789351683095441 1 +494 0 -0.08154678 0.4796246 0.9423753303896093 0 +496 0 -7.202647 0.000744058 0.0010738483301681407 0 +498 0 -4.9443603 0.0070730853 0.010240564107579979 0 +499 0 -4.9443603 0.0070730853 0.010240564107579979 0 +500 0 -3.322402 0.034810614 0.051116044045370665 0 +503 0 -4.5950603 0.010000589 0.014500428468588475 0 +505 0 -5.865587 0.002827341 0.0040847680370083737 0 +506 1 8.127695 0.99970484 0.00042589190022408524 1 +508 0 -4.6573305 0.00940252 0.01362914404396972 0 +509 0 -6.566318 0.0014049915 0.002028399530822185 0 +511 0 -4.308031 0.013281259 0.019289183899785744 0 +512 0 -4.6573305 0.00940252 0.01362914404396972 0 +515 1 5.338561 0.9952202 0.0069123494389670683 1 +516 0 -7.202647 0.000744058 0.0010738483301681407 0 +518 0 -5.298871 0.0049723843 0.0071915285634126186 0 +524 0 -5.5806894 0.0037558062 0.005428683982832385 0 +525 0 -5.6403227 0.003539152 0.0051149736903522642 0 +526 0 -5.29366 0.0049982327 0.0072290067512698264 0 +530 1 4.6355476 0.9903924 0.013927870202554189 1 +536 0 -3.9587312 0.018729815 0.027277669434812288 0 +537 0 -3.617279 0.026153287 0.038233390014612942 0 +542 0 -4.592929 0.010021714 0.014531212515523896 0 +543 0 -4.9443603 0.0070730853 0.010240564107579979 0 +545 0 -4.308031 0.013281259 0.019289183899785744 0 +550 0 -5.5806894 0.0037558062 0.005428683982832385 0 +551 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +552 0 -4.0065393 0.01787107 0.026015667577808772 0 +553 0 -3.0458336 0.04539769 0.067028269538105992 0 +554 0 -5.929989 0.0026514626 0.0038303321306625395 0 +555 0 -2.6957402 0.06322519 0.094225808725499016 0 +556 0 -3.6447124 0.025463587 0.037212002431777941 0 +562 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +564 0 -4.382829 0.012335897 0.017907619819922383 0 +567 0 -4.3053417 0.013316549 0.019340783703355929 0 +568 1 2.6288567 0.9326958 0.10052146471617009 1 +570 1 4.9677286 0.99308914 0.010004875333809272 1 +571 1 8.770559 0.9998448 0.00022393879069825088 1 +572 0 -5.5806894 0.0037558062 0.005428683982832385 0 +573 0 -6.8533473 0.0010548015 0.0015225600517844947 0 +574 1 5.0986586 0.9939321 0.0087807561877431065 1 +575 0 -3.617279 0.026153287 0.038233390014612942 0 +576 0 -4.308031 0.013281259 0.019289183899785744 0 +579 0 -6.5040474 0.0014951277 0.0021586274073604299 0 +580 0 -3.966579 0.018586125 0.027066427182224399 0 +583 0 -5.929989 0.0026514626 0.0038303321306625395 0 +585 0 -7.202647 0.000744058 0.0010738483301681407 0 +587 0 -4.0958 0.016369991 0.023812344938505647 0 +588 1 4.5078 0.9890975 0.015815388645031802 1 +589 0 -4.6573305 0.00940252 0.01362914404396972 0 +591 1 3.092064 0.9565642 0.064066314389843587 1 +592 1 4.695217 0.9909439 0.013124697357507648 1 +595 0 -4.308031 0.013281259 0.019289183899785744 0 +596 0 -4.602908 0.00992319 0.014387641126567057 0 +597 0 -3.2758274 0.036409825 0.053508411435479734 0 +598 0 -5.5806894 0.0037558062 0.005428683982832385 0 +599 0 -3.6007748 0.026576942 0.038861145718525385 0 +601 0 -6.8533473 0.0010548015 0.0015225600517844947 0 +603 1 3.5293732 0.9715121 0.041696161300407437 1 +605 1 8.488371 0.9997942 0.00029695860700686236 1 +608 1 7.3318195 0.999346 0.0009438054722544462 1 +610 1 7.0646324 0.9991459 0.001232696184539676 1 +611 1 4.92607 0.9927973 0.010428880977607462 1 +615 0 -4.2614565 0.013905655 0.020202410717548722 0 +616 0 -5.5806894 0.0037558062 0.005428683982832385 0 +620 0 -5.5806894 0.0037558062 0.005428683982832385 0 +623 0 -7.202647 0.000744058 0.0010738483301681407 0 +625 0 -4.288558 0.013538885 0.019665911857217805 0 +626 1 3.0648718 0.95542026 0.065792631484175745 1 +628 0 -6.566318 0.0014049915 0.002028399530822185 0 +630 0 -3.3058977 0.035369415 0.051951541096231811 0 +631 0 -4.308031 0.013281259 0.019289183899785744 0 +632 0 -7.202647 0.000744058 0.0010738483301681407 0 +635 0 -4.9056334 0.0073503237 0.010643440043832827 0 +636 1 8.309147 0.99975383 0.00035518789362715982 1 +637 0 -2.9889622 0.047927022 0.070855932076943823 0 +640 0 -4.714328 0.008886218 0.012877403568593912 0 +643 0 -7.202647 0.000744058 0.0010738483301681407 0 +646 0 -6.8289943 0.0010807763 0.0015600738517959583 0 +647 0 -6.623315 0.0013272542 0.001916094945843513 0 +648 1 8.327751 0.99975836 0.0003486509585214664 1 +650 0 -4.6817956 0.0091773635 0.013301265841350862 0 +651 0 -6.241814 0.0019425406 0.0028052192505183078 0 +655 0 -5.5806894 0.0037558062 0.005428683982832385 0 +658 1 6.517376 0.99852467 0.002130028925943152 1 +659 0 -7.202647 0.000744058 0.0010738483301681407 0 +662 0 -6.1704435 0.0020859484 0.003012530443726541 0 +663 0 -6.1704435 0.0020859484 0.003012530443726541 0 +664 0 -5.001357 0.006683835 0.0096751049020648225 0 +666 0 -3.811576 0.021634884 0.031555130237863965 0 +667 0 -5.2313895 0.005317663 0.0076922363466130801 0 +669 1 6.0171223 0.99756926 0.0035110812308474519 1 +671 0 -4.9095225 0.007322002 0.010602278762236021 0 +672 0 -5.8677187 0.0028213381 0.0040760832411949836 0 +673 0 -4.2514772 0.014043159 0.020403598815497921 0 +674 0 -6.8533473 0.0010548015 0.0015225600517844947 0 +675 0 -4.2693043 0.013798453 0.020045578844219699 0 +676 0 -6.511895 0.0014834575 0.0021417657985981662 0 +677 0 -4.6573305 0.00940252 0.01362914404396972 0 +684 0 -7.202647 0.000744058 0.0010738483301681407 0 +686 0 -7.202647 0.000744058 0.0010738483301681407 0 +687 0 -5.156227 0.005730377 0.0082909644737417 0 +690 0 -6.623315 0.0013272542 0.001916094945843513 0 +695 0 -6.566318 0.0014049915 0.002028399530822185 0 diff --git a/test/BaselineOutput/Common/LogisticRegression/netcoreapp/LogisticRegression-non-negative-TrainTest-breast-cancer.txt b/test/BaselineOutput/Common/LogisticRegression/netcoreapp/LogisticRegression-non-negative-TrainTest-breast-cancer.txt index a9187dbb6e..8fd5e1c875 100644 --- a/test/BaselineOutput/Common/LogisticRegression/netcoreapp/LogisticRegression-non-negative-TrainTest-breast-cancer.txt +++ b/test/BaselineOutput/Common/LogisticRegression/netcoreapp/LogisticRegression-non-negative-TrainTest-breast-cancer.txt @@ -1,700 +1,700 @@ Instance Label Score Probability Log-loss Assigned -0 0 -3.9978452 0.01802431 0.026240786261966513 0 -1 0 2.2570877 0.90526015 3.39988473602306 1 -2 0 -4.6860514 0.009138746 0.013245037929465322 0 -3 0 1.27454 0.78151894 2.1944198507861499 1 -4 0 -3.8983445 0.019872526 0.028958698138438609 0 -5 1 10.383755 0.99996907 4.4630188215668621E-05 1 -6 0 -2.6780577 0.06428061 0.095852140658507543 0 -7 0 -5.3104105 0.004915616 0.0071092222935852994 0 -8 0 -4.9831448 0.0068058423 0.0098523197281849221 0 -9 0 -4.9553556 0.0069962814 0.010128974585098 0 -10 0 -6.217933 0.0019893954 0.0028729494815213073 0 -11 0 -6.0393434 0.0023774575 0.0034340298239276387 0 -12 1 -0.8686228 0.29554096 1.7585700093704129 0 -13 0 -5.3742576 0.00461298 0.0066705206352207034 0 -14 1 8.05541 0.9996827 0.00045780445418641197 1 -15 1 0.7239399 0.673474 0.5703058149419834 1 -16 0 -4.9659657 0.0069229538 0.010022443686233343 0 -17 0 -4.534534 0.0106179565 0.015400378952762167 0 -18 1 6.5412884 0.9985595 0.002079736669315975 1 -19 0 -3.4611568 0.030437874 0.044594750482780984 0 -20 1 6.597271 0.9986378 0.0019665855015772657 1 -21 1 6.9094257 0.99900264 0.0014396108762453246 1 -22 0 -5.5026546 0.0040593916 0.0058683831829502422 0 +0 0 -3.9978604 0.01802404 0.026240389461776918 0 +1 0 2.2570457 0.90525657 3.3998302776097185 1 +2 0 -4.686054 0.00913872 0.013244999961279327 0 +3 0 1.2745934 0.78152806 2.1944800708672423 1 +4 0 -3.8983784 0.019871866 0.028957727572116825 0 +5 1 10.38364 0.99996907 4.4630188215668621E-05 1 +6 0 -2.6780891 0.064278714 0.095849222880120724 0 +7 0 -5.310356 0.004915882 0.0071096077905197456 0 +8 0 -4.983152 0.006805794 0.0098522493814660397 0 +9 0 -4.9553747 0.0069961487 0.010128781771069051 0 +10 0 -6.2179203 0.001989421 0.0028729865045729139 0 +11 0 -6.039318 0.0023775175 0.0034341166935918394 0 +12 1 -0.8686724 0.29553062 1.7586204922410928 0 +13 0 -5.3742485 0.0046130214 0.0066705807031576005 0 +14 1 8.055166 0.99968266 0.00045789047280600954 1 +15 1 0.72400665 0.6734887 0.57027440521369854 1 +16 0 -4.9659586 0.0069230027 0.010022514717742074 0 +17 0 -4.53454 0.010617891 0.015400283890384886 0 +18 1 6.541193 0.9985593 0.0020799950154694741 1 +19 0 -3.4611807 0.030437171 0.044593705593641481 0 +20 1 6.5969963 0.9986374 0.0019671882620641099 1 +21 1 6.9093056 0.9990025 0.0014397830306071015 1 +22 0 -5.502638 0.004059459 0.0058684809920473086 0 23 1 ? ? ? 0 -24 0 -6.1446 0.0021404426 0.0030913155089767924 0 -25 1 0.9981394 0.7306926 0.45266344873715486 1 -26 0 -5.565377 0.0038135375 0.0055122890455318622 0 -27 0 -4.429277 0.011782622 0.017099668747007298 0 -28 0 -6.0393434 0.0023774575 0.0034340298239276387 0 -29 0 -6.4124613 0.0016382922 0.0023654942741888757 0 -30 0 -5.5759873 0.003773441 0.0054542217988004307 0 -31 0 -5.6079116 0.0036553128 0.0052831633727484281 0 -32 1 6.2957153 0.9981592 0.0026581173462199756 1 -33 0 -5.289817 0.0050173826 0.0072567732008324446 0 -34 0 -5.2051535 0.005458252 0.0078961624282101964 0 -35 0 -6.0393434 0.0023774575 0.0034340298239276387 0 -36 1 7.935848 0.99964243 0.00051595421118705818 1 -37 0 -1.7981524 0.14207612 0.22107844541612048 0 -38 1 4.9257135 0.99279475 0.010432605435470237 1 -39 1 0.8958559 0.71009713 0.49391171066635314 1 +24 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +25 1 0.99808407 0.7306817 0.4526849851937591 1 +26 0 -5.565395 0.0038134705 0.0055121919349591939 0 +27 0 -4.4292784 0.011782605 0.017099644273586154 0 +28 0 -6.039318 0.0023775175 0.0034341166935918394 0 +29 0 -6.4123287 0.001638509 0.0023658075136661579 0 +30 0 -5.5759788 0.0037734734 0.005454268666255248 0 +31 0 -5.6078997 0.0036553564 0.005283226417172892 0 +32 1 6.295658 0.9981591 0.0026582896460455 1 +33 0 -5.289819 0.0050173732 0.0072567596969336824 0 +34 0 -5.2050943 0.005458573 0.0078966278438219947 0 +35 0 -6.039318 0.0023775175 0.0034341166935918394 0 +36 1 7.9357643 0.99964243 0.00051595421118705818 1 +37 0 -1.7982645 0.14206246 0.22105546744007815 0 +38 1 4.9255333 0.99279344 0.010434510975769298 1 +39 1 0.89574337 0.710074 0.4939586974460341 1 40 0 ? ? ? 0 -41 1 2.7223558 0.938333 0.091828129493380276 1 -42 1 6.456045 0.99843144 0.0022647243580549464 1 -43 1 -0.57765293 0.35947284 1.476045318936354 0 -44 1 7.0351467 0.9991204 0.0012695324025780532 1 -45 0 -6.2041245 0.0020170012 0.0029128561705458406 0 -46 1 2.966281 0.95102733 0.072441288313320681 1 -47 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -48 0 -3.8983445 0.019872526 0.028958698138438609 0 -49 1 4.042761 0.9827537 0.02509821330401326 1 -50 1 2.100398 0.89094186 0.16659680847266228 1 -51 1 -0.59678745 0.35507903 1.4937879538434611 0 -52 1 4.9548836 0.99300045 0.010133726613379825 1 -53 1 7.1513233 0.9992168 0.0011303687292114905 1 -54 1 6.3704414 0.9982915 0.002466963365520171 1 -55 1 3.674755 0.9752714 0.036124339189793395 1 -56 1 4.1845055 0.9849987 0.021806269985834032 1 -57 1 1.3889542 0.80042523 0.32116144947804109 1 -58 1 1.230504 0.77390677 0.36976831967664936 1 -59 1 0.947484 0.7206089 0.47271164467487931 1 -60 1 1.7859726 0.8564328 0.22358805274312279 1 -61 0 -6.1908607 0.002043877 0.002951708748520548 0 -62 1 4.6547976 0.9905739 0.013663510742586951 1 -63 1 -0.14216805 0.46451774 1.1061943930544347 0 -64 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -65 1 3.7020578 0.9759214 0.035163147204738461 1 -66 0 -4.534534 0.0106179565 0.015400378952762167 0 -67 1 3.0539446 0.95495254 0.06649906317790269 1 -68 1 10.140783 0.99996054 5.6927380571090205E-05 1 -69 0 -5.919588 0.0026791096 0.0038703249258569359 0 -70 0 -3.8342748 0.021159602 0.030854450849772463 0 -71 1 6.7783127 0.9988631 0.001641131618591557 1 -72 0 -3.259544 0.03698545 0.054370498494845383 0 -73 1 7.9999866 0.99966466 0.00048386833058238708 1 -74 1 1.828826 0.8616218 0.21487334761852855 1 -75 0 -4.8054843 0.00811829 0.011760017851163681 0 -76 0 -5.683529 0.0033900023 0.0048990480946461195 0 -77 0 -4.0899615 0.016464267 0.023950626529872041 0 -78 0 -4.30088 0.013375301 0.019426690787480647 0 -79 0 -5.96601 0.0025578933 0.0036949876647491787 0 -80 0 -3.647406 0.025396828 0.037113176290791992 0 -81 0 -4.647871 0.009491038 0.013758065775472702 0 -82 0 -3.987235 0.018213065 0.026518127030675635 0 -83 0 -2.9603248 0.049250796 0.072863269568515771 0 -84 1 8.486536 0.9997938 0.0002974746612633519 1 -85 1 5.131035 0.9941243 0.0085018546122276779 1 -86 1 1.4144535 0.80446744 0.31389406499389128 1 -87 1 5.359544 0.99531895 0.0067691845808085281 1 -88 0 -4.534534 0.0106179565 0.015400378952762167 0 -89 0 -5.6479154 0.003512476 0.0050763523697816629 0 -90 0 -6.1446 0.0021404426 0.0030913155089767924 0 -91 0 -5.6159916 0.0036260036 0.0052407246489976148 0 -92 0 -4.534534 0.0106179565 0.015400378952762167 0 -93 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -94 0 -5.6079116 0.0036553128 0.0052831633727484281 0 -95 0 -6.1446 0.0021404426 0.0030913155089767924 0 -96 0 -6.1526804 0.002123253 0.003066463099432681 0 -97 0 -3.9978452 0.01802431 0.026240786261966513 0 -98 1 8.128416 0.9997051 0.0004255478334078568 1 -99 1 9.796417 0.9999443 8.0318133658572531E-05 1 -100 1 4.4211864 0.9881228 0.017237719044512673 1 -101 1 -1.7306023 0.15051056 2.7320633405296841 0 -102 0 -4.2370334 0.014244557 0.020698323994567083 0 -103 1 1.0072336 0.73247844 0.44914179984395614 1 -104 1 11.24371 0.9999869 1.8918215632667518E-05 1 -105 1 2.2436733 0.9041034 0.14544031813461844 1 -106 1 7.5832367 0.99949133 0.00073403667682701737 1 +41 1 2.722229 0.93832564 0.091839401584024927 1 +42 1 6.4560623 0.9984315 0.0022646382316379881 1 +43 1 -0.5776367 0.35947657 1.476030368071148 0 +44 1 7.0351295 0.99912035 0.0012696184696093834 1 +45 0 -6.2041016 0.0020170473 0.002912922813882828 0 +46 1 2.9662647 0.9510266 0.072442373346585875 1 +47 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +48 0 -3.8983784 0.019871866 0.028957727572116825 0 +49 1 4.042717 0.982753 0.025099263309001214 1 +50 1 2.1004963 0.8909514 0.16658136578321556 1 +51 1 -0.5968437 0.35506612 1.4938403857218505 0 +52 1 4.9548893 0.9930005 0.010133640015913485 1 +53 1 7.151163 0.9992167 0.0011305408466758765 1 +54 1 6.370329 0.9982913 0.0024672217810244061 1 +55 1 3.674718 0.9752705 0.036125661765711949 1 +56 1 4.184555 0.9849995 0.021805135073892017 1 +57 1 1.3889618 0.8004264 0.32115930083859251 1 +58 1 1.2304993 0.77390593 0.36976987526351829 1 +59 1 0.9474516 0.7206024 0.47272465186484686 1 +60 1 1.785963 0.8564316 0.223590060872324 1 +61 0 -6.190832 0.0020439355 0.0029517932330062734 0 +62 1 4.6549053 0.99057484 0.013662121789609155 1 +63 1 -0.14213371 0.46452627 1.1061679211985749 0 +64 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +65 1 3.7018375 0.9759162 0.035170813052875949 1 +66 0 -4.53454 0.010617891 0.015400283890384886 0 +67 1 3.0539513 0.95495284 0.066498612939233273 1 +68 1 10.140743 0.99996054 5.6927380571090205E-05 1 +69 0 -5.919572 0.0026791529 0.0038703875717659309 0 +70 0 -3.8341913 0.02116133 0.030856998507843483 0 +71 1 6.7778645 0.99886256 0.0016419064216010202 1 +72 0 -3.2595644 0.03698472 0.054369404644979212 0 +73 1 7.999836 0.9996646 0.00048395435075601986 1 +74 1 1.8287258 0.8616099 0.21489330809784613 1 +75 0 -4.8054447 0.008118609 0.011760481128401914 0 +76 0 -5.683367 0.0033905501 0.0048998411645782697 0 +77 0 -4.0900307 0.016463147 0.023948984470913377 0 +78 0 -4.300889 0.013375181 0.019426515111506861 0 +79 0 -5.965977 0.0025579773 0.0036951092369279506 0 +80 0 -3.6474404 0.025395978 0.037111918983317373 0 +81 0 -4.6478777 0.009490974 0.013757973534227223 0 +82 0 -3.9872766 0.018212324 0.02651703767343476 0 +83 0 -2.960382 0.049248118 0.072859205163592913 0 +84 1 8.486427 0.99979377 0.00029756067032404669 1 +85 1 5.130885 0.9941234 0.0085031521063842468 1 +86 1 1.4145298 0.8044794 0.31387257981393474 1 +87 1 5.359432 0.9953185 0.0067698757469606191 1 +88 0 -4.53454 0.010617891 0.015400283890384886 0 +89 0 -5.6478963 0.003512543 0.0050764491139339138 0 +90 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +91 0 -5.6159754 0.003626062 0.0052408092676349298 0 +92 0 -4.53454 0.010617891 0.015400283890384886 0 +93 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +94 0 -5.6078997 0.0036553564 0.005283226417172892 0 +95 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +96 0 -6.152655 0.0021233065 0.0030665405216531542 0 +97 0 -3.9978604 0.01802404 0.026240389461776918 0 +98 1 8.128468 0.9997051 0.0004255478334078568 1 +99 1 9.796211 0.9999443 8.0318133658572531E-05 1 +100 1 4.421232 0.98812336 0.017236935820299577 1 +101 1 -1.7305841 0.15051287 2.7320412016299604 0 +102 0 -4.2369967 0.014245073 0.020699079113507712 0 +103 1 1.0072165 0.73247504 0.44914849153158898 1 +104 1 11.243553 0.9999869 1.8918215632667518E-05 1 +105 1 2.2435951 0.9040966 0.14545116097400559 1 +106 1 7.5831585 0.99949133 0.00073403667682701737 1 107 1 4.9732876 0.9931272 0.0099495455541438165 1 -108 0 -6.126008 0.002180522 0.0031492631521246343 0 -109 1 5.6274014 0.99641496 0.0051814135991893788 1 -110 0 -3.7353187 0.023309276 0.034026299202021874 0 -111 1 2.4953337 0.92381406 0.11432559406639971 1 -112 1 8.550783 0.99980664 0.00027898283227574002 1 -113 1 7.534811 0.9994661 0.00077042997831079868 1 -114 0 -3.9821587 0.018304061 0.02665184777383104 0 -115 0 -5.4667783 0.0042070476 0.006082289989879185 0 -116 0 -1.1613731 0.23841786 0.39292845603147281 0 -117 1 8.403101 0.9997759 0.00032336362004626066 1 -118 0 -6.090455 0.0022592624 0.0032631141249073411 0 -119 0 -4.6754413 0.009235324 0.013385662673546376 0 -120 0 -5.529004 0.003954243 0.005716075482133186 0 -121 0 -4.1387525 0.015692547 0.022819076386462446 0 -122 1 7.8866005 0.9996244 0.00054201913815316223 1 -123 1 2.432806 0.91929495 0.12140027422407701 1 -124 1 6.4366856 0.99840087 0.0023089078878419595 1 -125 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -126 1 7.348158 0.9993566 0.00092848908093139148 1 -127 0 -5.071223 0.006235616 0.009024257247501357 0 -128 1 4.0983744 0.9836714 0.023751598556500764 1 -129 0 -6.4208 0.0016247095 0.0023458664742784341 0 -130 0 -3.8342748 0.021159602 0.030854450849772463 0 -131 0 -5.6079116 0.0036553128 0.0052831633727484281 0 -132 1 6.8184223 0.99890774 0.0015766522486064089 1 -133 0 -5.2776427 0.005078526 0.0073454323967635073 0 -134 0 -5.429322 0.004366916 0.0063139245024787915 0 -135 0 -3.16265 0.040595714 0.059789210306242166 0 -136 0 -4.9659657 0.0069229538 0.010022443686233343 0 -137 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -138 0 -4.668465 0.009299376 0.013478934137374143 0 +108 0 -6.1259823 0.0021805782 0.003149344281803172 0 +109 1 5.6272116 0.9964143 0.0051823629073887391 1 +110 0 -3.7354183 0.023307007 0.034022948048027295 0 +111 1 2.4952374 0.92380726 0.11433620556019902 1 +112 1 8.550491 0.9998066 0.00027906884023401566 1 +113 1 7.5348206 0.9994661 0.00077042997831079868 1 +114 0 -3.9821777 0.018303718 0.026651344104603791 0 +115 0 -5.4667535 0.0042071515 0.0060824404358325633 0 +116 0 -1.1613932 0.23841423 0.39292156845089399 0 +117 1 8.402843 0.9997758 0.00032344963065038983 1 +118 0 -6.090403 0.0022593795 0.0032632834670238572 0 +119 0 -4.6754704 0.009235059 0.013385276174028201 0 +120 0 -5.528955 0.003954436 0.0057163553889803213 0 +121 0 -4.138791 0.015691949 0.022818200034039646 0 +122 1 7.8866253 0.99962443 0.00054193311451736768 1 +123 1 2.432867 0.9192995 0.12139316516228225 1 +124 1 6.4366302 0.99840075 0.0023090801459667441 1 +125 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +126 1 7.347864 0.99935645 0.00092874722101100274 1 +127 0 -5.07122 0.0062356335 0.009024282936362189 0 +128 1 4.0983734 0.98367137 0.023751685975252283 1 +129 0 -6.42069 0.0016248882 0.0023461246998675508 0 +130 0 -3.8341913 0.02116133 0.030856998507843483 0 +131 0 -5.6078997 0.0036553564 0.005283226417172892 0 +132 1 6.8182907 0.9989076 0.0015768244193218785 1 +133 0 -5.277631 0.0050785863 0.0073455195022663648 0 +134 0 -5.4292974 0.004367022 0.0063140776716048095 0 +135 0 -3.1626797 0.04059456 0.059787473727667503 0 +136 0 -4.9659586 0.0069230027 0.010022514717742074 0 +137 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +138 0 -4.668415 0.009299837 0.013479605469657881 0 139 0 ? ? ? 0 -140 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -141 0 -6.470775 0.0015456327 0.0022316017284886259 0 -142 1 3.648261 0.97462434 0.037081848706034023 1 -143 0 -5.4667783 0.0042070476 0.006082289989879185 0 -144 0 -6.0393434 0.0023774575 0.0034340298239276387 0 +140 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +141 0 -6.470736 0.001545693 0.0022316888622042515 0 +142 1 3.64818 0.9746223 0.037084848536837592 1 +143 0 -5.4667535 0.0042071515 0.0060824404358325633 0 +144 0 -6.039318 0.0023775175 0.0034341166935918394 0 145 0 ? ? ? 0 -146 1 -0.11033058 0.4724453 1.0817807640029504 0 -147 0 -6.1175275 0.002199052 0.0031760547091254222 0 -148 0 -1.5570889 0.17406477 0.27589944582558007 0 -149 1 10.011244 0.9999551 6.475292119065462E-05 1 -150 0 -6.217933 0.0019893954 0.0028729494815213073 0 -151 1 4.020233 0.9823677 0.025664977101845805 1 -152 1 8.3236885 0.99975735 0.00035011316511816354 1 -153 0 -4.6270547 0.009688742 0.014046055417410509 0 -154 0 -7.0074635 0.000904283 0.0013051948511978793 0 -155 1 2.3649483 0.9141151 0.12955230778828825 1 -156 0 -6.2679205 0.0018925789 0.0027330012268123789 0 -157 0 -5.6079116 0.0036553128 0.0052831633727484281 0 +146 1 -0.11027336 0.47245955 1.0817372634822817 0 +147 0 -6.1174912 0.0021991315 0.0031761698413480576 0 +148 0 -1.5571251 0.17405957 0.27589036191772137 0 +149 1 10.011087 0.9999551 6.475292119065462E-05 1 +150 0 -6.2179203 0.001989421 0.0028729865045729139 0 +151 1 4.020218 0.98236746 0.025665327240948946 1 +152 1 8.3235035 0.9997573 0.00035019917731706222 1 +153 0 -4.627045 0.009688834 0.014046188379872699 0 +154 0 -7.007416 0.0009043261 0.0013052570496129583 0 +155 1 2.3648844 0.91411006 0.12956020973777446 1 +156 0 -6.2678704 0.0018926736 0.0027331380305502051 0 +157 0 -5.6078997 0.0036553564 0.005283226417172892 0 158 0 ? ? ? 0 -159 1 10.917734 0.9999819 2.614159977229895E-05 1 -160 1 7.7599354 0.9995737 0.00061514108379339587 1 -161 0 -4.309522 0.013261733 0.019260634789626564 0 -162 0 -5.071223 0.006235616 0.009024257247501357 0 -163 0 -3.7609138 0.022733632 0.033176252518656739 0 +159 1 10.917658 0.9999819 2.614159977229895E-05 1 +160 1 7.7598343 0.99957365 0.00061522711179447589 1 +161 0 -4.3095326 0.013261596 0.019260434623768575 0 +162 0 -5.07122 0.0062356335 0.009024282936362189 0 +163 0 -3.7609682 0.022732424 0.033174467938169562 0 164 0 ? ? ? 0 -165 0 -3.9243507 0.01937226 0.028222522016834069 0 -166 1 6.742338 0.9988215 0.0017012231317571679 1 -167 1 7.940733 0.99964416 0.0005134595729021004 1 -168 0 -5.071223 0.006235616 0.009024257247501357 0 -169 0 -6.762702 0.0011547662 0.0016669380960040003 0 -170 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -171 0 -6.1446 0.0021404426 0.0030913155089767924 0 -172 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -173 1 13.357813 0.9999984 2.3217676548882582E-06 1 -174 1 4.643731 0.99047 0.013814827815610272 1 -175 1 6.0361423 0.9976149 0.0034450528853451077 1 -176 0 -5.6079116 0.0036553128 0.0052831633727484281 0 -177 1 3.4831276 0.9702039 0.043640150987015548 1 -178 0 -4.534534 0.0106179565 0.015400378952762167 0 -179 1 1.0144405 0.73388827 0.44636765960242319 1 -180 0 -6.217933 0.0019893954 0.0028729494815213073 0 -181 0 -7.0074635 0.000904283 0.0013051948511978793 0 -182 0 -3.4611568 0.030437874 0.044594750482780984 0 -183 1 7.3729496 0.99937236 0.00090577293479888533 1 -184 1 4.5937433 0.98998636 0.014519446857209574 1 -185 0 -5.6812444 0.0033977295 0.0049102340273752066 0 -186 1 4.237855 0.98576695 0.020681486580487643 1 -187 1 13.26124 0.9999983 2.4937505927785606E-06 1 -188 1 7.361681 0.99936527 0.00091601236548379417 1 -189 0 -5.538988 0.003915114 0.0056594011973182395 0 -190 1 10.425629 0.9999703 4.2824315647486083E-05 1 -191 1 9.911894 0.9999504 7.1546556808546518E-05 1 -192 0 -4.429277 0.011782622 0.017099668747007298 0 -193 0 -6.1446 0.0021404426 0.0030913155089767924 0 -194 0 -5.071223 0.006235616 0.009024257247501357 0 -195 0 -4.534534 0.0106179565 0.015400378952762167 0 -196 0 5.9393463 0.99737316 8.5724579877636575 1 -197 0 -3.04356 0.045496322 0.067177338418410296 0 -198 0 -7.0074635 0.000904283 0.0013051948511978793 0 -199 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -200 1 9.001123 0.99987674 0.00017784102177338968 1 -201 1 8.007027 0.999667 0.0004805135478112585 1 -202 0 -6.1446 0.0021404426 0.0030913155089767924 0 -203 0 -3.9978452 0.01802431 0.026240786261966513 0 -204 0 -6.1446 0.0021404426 0.0030913155089767924 0 -205 1 10.318253 0.999967 4.7639980852592516E-05 1 -206 1 4.604685 0.99009424 0.014362236793309195 1 -207 0 -6.217933 0.0019893954 0.0028729494815213073 0 -208 0 -6.217933 0.0019893954 0.0028729494815213073 0 -209 0 -4.071178 0.016771212 0.024400938148352504 0 -210 1 12.976419 0.9999977 3.3536655897601792E-06 1 -211 1 7.935549 0.9996423 0.00051612625536571392 1 -212 0 -6.1446 0.0021404426 0.0030913155089767924 0 -213 1 13.021215 0.9999978 3.1816825493598248E-06 1 -214 1 12.006409 0.9999939 8.7711418568707846E-06 1 -215 1 6.687521 0.99875516 0.0017970475849021653 1 -216 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -217 0 -6.1446 0.0021404426 0.0030913155089767924 0 -218 1 6.330929 0.99822277 0.0025662844675125853 1 -219 0 -2.8661547 0.053852245 0.079862595830934621 0 -220 0 -5.826505 0.0029396957 0.0042473304372818805 0 -221 1 8.934246 0.9998682 0.00019013934964380927 1 -222 1 -2.8624234 0.054042675 4.2097571135895304 0 -223 1 4.4119625 0.98801404 0.017396548295654901 1 -224 1 8.946123 0.99986976 0.00018790328223314414 1 -225 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -226 1 9.037792 0.9998812 0.0001713908917246428 1 -227 1 6.0393066 0.99762243 0.0034341921153200391 1 -228 0 -6.217933 0.0019893954 0.0028729494815213073 0 -229 1 10.761692 0.9999788 3.0613236649743679E-05 1 -230 1 5.0818014 0.9938296 0.0089295718905516644 1 -231 1 6.989193 0.99907905 0.001329264157056246 1 -232 0 0.7225342 0.67316484 1.6133649219331099 1 -233 1 5.2977858 0.99502224 0.0071993267333871911 1 -234 0 -3.554109 0.027811257 0.040691665919696543 0 +165 0 -3.924367 0.019371953 0.02822206986492741 0 +166 1 6.742281 0.99882144 0.001701309224545778 1 +167 1 7.94067 0.99964416 0.0005134595729021004 1 +168 0 -5.07122 0.0062356335 0.009024282936362189 0 +169 0 -6.762676 0.0011547965 0.0016669818139586235 0 +170 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +171 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +172 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +173 1 13.357599 0.9999984 2.3217676548882582E-06 1 +174 1 4.6436005 0.99046874 0.01381665100963876 1 +175 1 6.03607 0.99761474 0.0034453114761041339 1 +176 0 -5.6078997 0.0036553564 0.005283226417172892 0 +177 1 3.4831886 0.9702056 0.043637580654865296 1 +178 0 -4.53454 0.010617891 0.015400283890384886 0 +179 1 1.0144091 0.7338821 0.44637972839265566 1 +180 0 -6.2179203 0.001989421 0.0028729865045729139 0 +181 0 -7.007416 0.0009043261 0.0013052570496129583 0 +182 0 -3.4611807 0.030437171 0.044593705593641481 0 +183 1 7.372835 0.9993723 0.00090585898013210812 1 +184 1 4.593649 0.9899854 0.014520836635821776 1 +185 0 -5.6812406 0.0033977423 0.0049102525650601874 0 +186 1 4.2378864 0.9857674 0.020680788717327418 1 +187 1 13.260954 0.9999983 2.4937505927785606E-06 1 +188 1 7.3613434 0.99936503 0.0009163565492902898 1 +189 0 -5.5389085 0.0039154245 0.0056598510540507973 0 +190 1 10.425512 0.9999703 4.2824315647486083E-05 1 +191 1 9.911949 0.9999504 7.1546556808546518E-05 1 +192 0 -4.4292784 0.011782605 0.017099644273586154 0 +193 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +194 0 -5.07122 0.0062356335 0.009024282936362189 0 +195 0 -4.53454 0.010617891 0.015400283890384886 0 +196 0 5.9391804 0.99737275 8.5722288560592155 1 +197 0 -3.0436182 0.045493796 0.06717352085575562 0 +198 0 -7.007416 0.0009043261 0.0013052570496129583 0 +199 0 -5.502638 0.004059459 0.0058684809920473086 0 +200 1 9.000977 0.99987674 0.00017784102177338968 1 +201 1 8.006981 0.999667 0.0004805135478112585 1 +202 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +203 0 -3.9978604 0.01802404 0.026240389461776918 0 +204 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +205 1 10.3181095 0.999967 4.7639980852592516E-05 1 +206 1 4.604699 0.99009436 0.014362063090006235 1 +207 0 -6.2179203 0.001989421 0.0028729865045729139 0 +208 0 -6.2179203 0.001989421 0.0028729865045729139 0 +209 0 -4.0712013 0.016770827 0.024400372403854145 0 +210 1 12.976204 0.9999977 3.3536655897601792E-06 1 +211 1 7.935364 0.99964225 0.00051621227746273556 1 +212 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +213 1 13.021034 0.9999978 3.1816825493598248E-06 1 +214 1 12.006208 0.9999939 8.7711418568707846E-06 1 +215 1 6.687456 0.9987551 0.0017971336834092871 1 +216 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +217 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +218 1 6.3307343 0.9982224 0.0025668013341493649 1 +219 0 -2.8660932 0.05385538 0.079867373019807619 0 +220 0 -5.826499 0.002939714 0.0042473567150126193 0 +221 1 8.934042 0.9998682 0.00019013934964380927 1 +222 1 -2.8624754 0.05404002 4.2098280220454809 0 +223 1 4.411908 0.9880134 0.01739750567566473 1 +224 1 8.945911 0.99986976 0.00018790328223314414 1 +225 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +226 1 9.037701 0.99988115 0.0001714768932689376 1 +227 1 6.039242 0.9976223 0.0034343645078564445 1 +228 0 -6.2179203 0.001989421 0.0028729865045729139 0 +229 1 10.761639 0.9999788 3.0613236649743679E-05 1 +230 1 5.0817995 0.9938296 0.0089295718905516644 1 +231 1 6.9890594 0.9990789 0.0013294362982510378 1 +232 0 0.7225838 0.67317575 1.6134130705942484 1 +233 1 5.297718 0.9950219 0.0071998452625476137 1 +234 0 -3.5541306 0.027810676 0.040690803520107427 0 235 0 ? ? ? 0 -236 1 10.420694 0.9999702 4.2996303413732479E-05 1 -237 1 5.8288326 0.9970671 0.004237506305358023 1 -238 1 11.358864 0.9999883 1.6854398235588073E-05 1 -239 1 4.855196 0.9922724 0.011191902672743228 1 -240 0 -2.3628898 0.086046666 0.12980759086355939 0 -241 0 -4.508366 0.010896405 0.015806463822240383 0 -242 0 -5.6079116 0.0036553128 0.0052831633727484281 0 -243 0 -3.7410512 0.023179125 0.033834063400741896 0 -244 0 -6.1446 0.0021404426 0.0030913155089767924 0 -245 0 -3.294579 0.03575763 0.05253227155054211 0 -246 1 9.599237 0.9999322 9.7775450197580527E-05 1 -247 1 2.0679493 0.8877488 0.17177662574908537 1 -248 0 -3.4094744 0.032000676 0.04692205458789344 0 +236 1 10.420582 0.9999702 4.2996303413732479E-05 1 +237 1 5.8287754 0.997067 0.0042376787939119084 1 +238 1 11.358711 0.9999883 1.6854398235588073E-05 1 +239 1 4.855036 0.9922711 0.01119372255508084 1 +240 0 -2.3628492 0.086049855 0.12981262453843617 0 +241 0 -4.508399 0.010896051 0.015805947624133235 0 +242 0 -5.6078997 0.0036553564 0.005283226417172892 0 +243 0 -3.7410808 0.023178456 0.033833075794030457 0 +244 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +245 0 -3.2946138 0.035756428 0.052530471226417248 0 +246 1 9.599129 0.9999322 9.7775450197580527E-05 1 +247 1 2.0679007 0.88774395 0.17178447179345496 1 +248 0 -3.4095025 0.031999804 0.046920755390118481 0 249 0 ? ? ? 0 -250 0 -6.6993523 0.0012301942 0.0017758876872952787 0 -251 1 7.996031 0.9996633 0.00048584679587356108 1 -252 0 3.6111832 0.973691 5.248299043804562 1 -253 1 6.456045 0.99843144 0.0022647243580549464 1 -254 1 4.6547976 0.9905739 0.013663510742586951 1 -255 1 2.5990734 0.9308019 0.10345389577310687 1 -256 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -257 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -258 0 -5.071223 0.006235616 0.009024257247501357 0 -259 0 1.4420557 0.8087728 2.3866403669656027 1 -260 1 8.819954 0.99985224 0.00021318824647088272 1 -261 1 10.051548 0.9999569 6.2173067944539402E-05 1 -262 1 8.008137 0.99966735 0.0004799974280773332 1 -263 1 7.995571 0.9996632 0.00048601883646188653 1 -264 1 4.3294783 0.9869969 0.018882557385354833 1 -265 0 -2.8803306 0.053134505 0.078768593220937502 0 -266 1 5.698023 0.9966586 0.0048286593976634177 1 -267 1 1.7236986 0.8486046 0.23683556271645489 1 -268 1 8.499901 0.99979657 0.00029351825001431117 1 -269 0 -6.1446 0.0021404426 0.0030913155089767924 0 -270 1 5.140936 0.9941818 0.0084183849403797993 1 -271 0 -3.9978452 0.01802431 0.026240786261966513 0 -272 1 1.7236986 0.8486046 0.23683556271645489 1 -273 1 -0.34248447 0.4152061 1.2681004399882052 0 -274 0 -4.846211 0.0077968277 0.011292525029475075 0 +250 0 -6.6992884 0.0012302727 0.0017760010262468937 0 +251 1 7.995781 0.99966323 0.00048593281616515928 1 +252 0 3.6110506 0.9736876 5.2481127506499741 1 +253 1 6.4560623 0.9984315 0.0022646382316379881 1 +254 1 4.6549053 0.99057484 0.013662121789609155 1 +255 1 2.5990658 0.93080145 0.10345463484632851 1 +256 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +257 0 -5.502638 0.004059459 0.0058684809920473086 0 +258 0 -5.07122 0.0062356335 0.009024282936362189 0 +259 0 1.4421072 0.8087807 2.3867001758394957 1 +260 1 8.819866 0.99985224 0.00021318824647088272 1 +261 1 10.051455 0.9999569 6.2173067944539402E-05 1 +262 1 8.008095 0.99966735 0.0004799974280773332 1 +263 1 7.995367 0.9996631 0.00048610485676374259 1 +264 1 4.3292274 0.9869937 0.018887262100431373 1 +265 0 -2.8802218 0.053139973 0.078776925689415411 0 +266 1 5.697832 0.99665797 0.0048296084737751788 1 +267 1 1.7237072 0.84860575 0.23683363739804933 1 +268 1 8.499756 0.9997965 0.00029360425883913718 1 +269 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +270 1 5.14098 0.9941821 0.0084179524676091192 1 +271 0 -3.9978604 0.01802404 0.026240389461776918 0 +272 1 1.7237072 0.84860575 0.23683363739804933 1 +273 1 -0.34243393 0.41521835 1.2680578805062157 0 +274 0 -4.8462124 0.0077968165 0.011292508779402632 0 275 0 ? ? ? 0 -276 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -277 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -278 0 -6.1446 0.0021404426 0.0030913155089767924 0 -279 1 6.586973 0.99862367 0.0019869933895667402 1 -280 0 -5.071223 0.006235616 0.009024257247501357 0 -281 0 -5.289817 0.0050173826 0.0072567732008324446 0 -282 1 2.9462185 0.95008445 0.073872342396203369 1 -283 1 4.9843636 0.9932024 0.0098403642172744926 1 -284 1 6.5737534 0.99860543 0.0020133432415614973 1 -285 1 12.268524 0.9999953 6.7933307031989985E-06 1 -286 1 14.178364 0.9999993 1.031896274211761E-06 1 -287 0 -5.429322 0.004366916 0.0063139245024787915 0 -288 1 1.7091665 0.84672815 0.2400292484255111 1 -289 1 7.101511 0.9991768 0.0011881152907599552 1 -290 0 -7.0074635 0.000904283 0.0013051948511978793 0 -291 0 -6.1446 0.0021404426 0.0030913155089767924 0 +276 0 -5.502638 0.004059459 0.0058684809920473086 0 +277 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +278 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +279 1 6.586918 0.9986236 0.0019870794994103725 1 +280 0 -5.07122 0.0062356335 0.009024282936362189 0 +281 0 -5.289819 0.0050173732 0.0072567596969336824 0 +282 1 2.9460316 0.95007557 0.07388582832095783 1 +283 1 4.98435 0.9932023 0.0098404507971388035 1 +284 1 6.5735893 0.9986052 0.0020136876872578749 1 +285 1 12.2683525 0.9999953 6.7933307031989985E-06 1 +286 1 14.178179 0.9999993 1.031896274211761E-06 1 +287 0 -5.4292974 0.004367022 0.0063140776716048095 0 +288 1 1.7091541 0.84672654 0.24003199047206264 1 +289 1 7.10137 0.9991767 0.0011882874151138022 1 +290 0 -7.007416 0.0009043261 0.0013052570496129583 0 +291 0 -6.1445794 0.0021404864 0.0030913787943164287 0 292 1 ? ? ? 0 -293 1 4.685541 0.99085665 0.01325174475399785 1 +293 1 4.6854115 0.99085546 0.013253480451636391 1 294 0 ? ? ? 0 -295 1 6.2169037 0.99800855 0.0028759207527209812 1 -296 0 1.0209227 0.7351523 1.9167651376310417 1 +295 1 6.2168503 0.99800843 0.0028760930785605436 1 +296 0 1.0209064 0.73514915 1.9167479295784922 1 297 0 ? ? ? 0 -298 0 -3.0200338 0.046528973 0.068738994193599606 0 -299 1 7.0977993 0.99917376 0.0011925044681979173 1 -300 1 6.6469555 0.9987037 0.0018713525078078205 1 -301 0 -6.1446 0.0021404426 0.0030913155089767924 0 -302 1 14.445042 0.99999946 7.7392213646475052E-07 1 -303 0 -6.1446 0.0021404426 0.0030913155089767924 0 -304 1 5.060603 0.99369824 0.0091202860803278971 1 -305 1 7.360697 0.9993647 0.0009168728251539932 1 -306 0 -6.1446 0.0021404426 0.0030913155089767924 0 -307 0 -6.1446 0.0021404426 0.0030913155089767924 0 -308 1 6.2902565 0.99814916 0.0026726767540725645 1 -309 0 -2.6676736 0.06490803 0.09681982450824847 0 -310 0 -5.96601 0.0025578933 0.0036949876647491787 0 -311 0 -7.0074635 0.000904283 0.0013051948511978793 0 -312 1 2.907339 0.94820803 0.076724478011134481 1 -313 0 -7.0074635 0.000904283 0.0013051948511978793 0 -314 0 -6.6493645 0.0012931703 0.0018668576723168487 0 +298 0 -3.0200515 0.04652819 0.068737810480984401 0 +299 1 7.0975914 0.9991736 0.0011927626555218829 1 +300 1 6.647064 0.99870384 0.0018711803019398868 1 +301 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +302 1 14.444851 0.99999946 7.7392213646475052E-07 1 +303 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +304 1 5.0604706 0.9936974 0.0091214975940590061 1 +305 1 7.360426 0.9993645 0.00091713096315512678 1 +306 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +307 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +308 1 6.2900963 0.99814886 0.0026731075080220256 1 +309 0 -2.6677213 0.06490514 0.096815364441260293 0 +310 0 -5.965977 0.0025579773 0.0036951092369279506 0 +311 0 -7.007416 0.0009043261 0.0013052570496129583 0 +312 1 2.9072113 0.9482018 0.076734000308609079 1 +313 0 -7.007416 0.0009043261 0.0013052570496129583 0 +314 0 -6.6493382 0.0012932041 0.001866906609577866 0 315 0 ? ? ? 0 -316 1 2.1206913 0.8928981 0.16343258239261421 1 -317 1 7.5715694 0.9994854 0.00074264021133653045 1 -318 0 -5.9246016 0.0026657472 0.0038509954247828359 0 -319 0 1.3538532 0.79475886 2.284608124824651 1 -320 1 5.9885216 0.9974989 0.003612801828101682 1 +316 1 2.120675 0.89289653 0.16343508634737081 1 +317 1 7.5714073 0.9994853 0.00074272624694072793 1 +318 0 -5.9245577 0.0026658638 0.0038511641623153793 0 +319 0 1.3537979 0.79474986 2.2845448606798033 1 +320 1 5.9882736 0.9974983 0.0036137501046989199 1 321 0 ? ? ? 0 -322 0 -5.071223 0.006235616 0.009024257247501357 0 -323 1 4.282318 0.98637754 0.019788149228401844 1 -324 0 -6.1446 0.0021404426 0.0030913155089767924 0 -325 0 -4.7445197 0.008624215 0.01249607394875411 0 -326 1 3.6900406 0.9756374 0.035583066579071274 1 -327 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -328 1 2.0947618 0.890393 0.16748581385828051 1 -329 1 6.5714874 0.9986023 0.0020179071537125624 1 -330 1 4.8110237 0.9919262 0.011695317633988674 1 -331 0 -3.6259375 0.025933666 0.037908071808741048 0 -332 0 -3.5780592 0.027170971 0.039741815820484569 0 -333 1 3.6066284 0.9735741 0.038637303501152256 1 -334 1 4.6732855 0.99074495 0.01341438869248103 1 -335 0 -7.0074635 0.000904283 0.0013051948511978793 0 -336 1 4.2861967 0.9864296 0.01971204404126722 1 -337 0 -6.1446 0.0021404426 0.0030913155089767924 0 -338 0 -6.6493645 0.0012931703 0.0018668576723168487 0 -339 1 4.465044 0.98862666 0.016502284739810944 1 -340 1 5.61915 0.9963854 0.0052242193898557636 1 -341 0 -6.1446 0.0021404426 0.0030913155089767924 0 -342 0 -6.470775 0.0015456327 0.0022316017284886259 0 -343 0 -7.0074635 0.000904283 0.0013051948511978793 0 -344 1 8.533409 0.99980325 0.00028388529408090524 1 -345 0 -7.0074635 0.000904283 0.0013051948511978793 0 -346 0 -3.7111464 0.023865968 0.034848838450133222 0 -347 0 -6.6307726 0.0013174057 0.0019018677457808206 0 -348 1 -0.61401176 0.3511446 1.5098627970874123 0 -349 1 2.9637775 0.9509106 0.072618340371597367 1 -350 0 -4.224519 0.014421356 0.020957099406612927 0 -351 0 -5.6079116 0.0036553128 0.0052831633727484281 0 -352 0 -0.51970005 0.37292236 0.67328401930104997 0 -353 1 7.176571 0.9992363 0.0011022277999205644 1 -354 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -355 0 -4.667621 0.009307154 0.013490261385739104 0 -356 1 -1.3441086 0.20683521 2.2734462930117343 0 -357 1 11.126102 0.9999853 2.1240013733369503E-05 1 -358 1 4.329771 0.98700064 0.018877068570489829 1 -359 1 4.8793297 0.99245524 0.010926051193307739 1 -360 1 13.757338 0.9999989 1.5478446880940214E-06 1 -361 1 4.220587 0.9855226 0.021039098622067324 1 -362 0 -4.277535 0.013686896 0.01988239407805973 0 -363 0 -2.9531941 0.049585763 0.073371647880799432 0 -364 0 -5.6079116 0.0036553128 0.0052831633727484281 0 -365 0 -6.0393434 0.0023774575 0.0034340298239276387 0 -366 1 12.063517 0.9999942 8.3411826797711399E-06 1 -367 1 9.634409 0.99993455 9.4421565109537245E-05 1 -368 0 -6.4124613 0.0016382922 0.0023654942741888757 0 -369 0 -6.054363 0.0023420993 0.0033828981697738799 0 -370 0 -4.5739117 0.010212158 0.014808773617867744 0 -371 0 -6.4124613 0.0016382922 0.0023654942741888757 0 -372 0 -4.668465 0.009299376 0.013478934137374143 0 -373 0 -4.111182 0.01612414 0.023451799790828513 0 -374 0 -5.2051535 0.005458252 0.0078961624282101964 0 -375 0 -7.0074635 0.000904283 0.0013051948511978793 0 -376 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -377 0 -6.6493645 0.0012931703 0.0018668576723168487 0 -378 0 -4.250025 0.014063283 0.020433045456902974 0 -379 0 -2.9374561 0.05033273 0.074505960636712887 0 -380 0 -7.0074635 0.000904283 0.0013051948511978793 0 -381 1 8.953867 0.9998708 0.00018644124003079258 1 -382 0 -4.220005 0.014485653 0.02105122023755461 0 -383 0 -6.470775 0.0015456327 0.0022316017284886259 0 -384 0 -6.470775 0.0015456327 0.0022316017284886259 0 -385 0 -4.25511 0.013992951 0.020330134736446452 0 -386 1 4.4445467 0.98839384 0.01684207091636555 1 -387 0 -2.9901948 0.04787081 0.070770757093141207 0 -388 0 -5.8143315 0.002975594 0.0042992743601272827 0 -389 0 -3.745761 0.023072727 0.033676929646663196 0 -390 0 -6.257937 0.0019115319 0.0027603967285926767 0 -391 1 8.5648575 0.9998093 0.0002751124794602528 1 -392 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -393 0 -7.0807962 0.00084039633 0.001212945365489941 0 -394 0 -5.97042 0.0025466671 0.0036787502937291913 0 -395 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -396 0 -5.071223 0.006235616 0.009024257247501357 0 -397 0 -5.3973975 0.004507937 0.0065182814231615977 0 -398 0 -5.2670326 0.0051324186 0.0074235816517891585 0 -399 0 -6.4018517 0.0016557375 0.0023907040452158406 0 -400 1 8.032017 0.9996752 0.00046864284064470924 1 -401 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -402 0 -3.8130546 0.021603609 0.031509011934098975 0 -403 0 -4.778572 0.008337892 0.012079463854828196 0 -404 0 -6.1356516 0.0021596404 0.0031190717827871335 0 -405 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -406 0 -4.657855 0.009397636 0.013622031269102946 0 -407 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -408 0 -4.5859237 0.010091454 0.014632848508026779 0 -409 0 -5.2051535 0.005458252 0.0078961624282101964 0 -410 0 -6.5760317 0.0013914289 0.0020088054126509622 0 +322 0 -5.07122 0.0062356335 0.009024282936362189 0 +323 1 4.2822437 0.9863765 0.019789631270751742 1 +324 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +325 0 -4.7444954 0.008624423 0.012496377536631818 0 +326 1 3.689888 0.97563374 0.035588443044500748 1 +327 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +328 1 2.0945282 0.8903702 0.16752280325415086 1 +329 1 6.571434 0.99860215 0.0020180793770952932 1 +330 1 4.8110456 0.9919264 0.011695057560250527 1 +331 0 -3.6259646 0.025932979 0.037907053821477674 0 +332 0 -3.578063 0.02717087 0.039741666657219656 0 +333 1 3.606617 0.9735738 0.038637745128237561 1 +334 1 4.673271 0.99074477 0.01341464907634672 1 +335 0 -7.007416 0.0009043261 0.0013052570496129583 0 +336 1 4.286089 0.9864281 0.019714223400869087 1 +337 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +338 0 -6.6493382 0.0012932041 0.001866906609577866 0 +339 1 4.4649353 0.9886254 0.016504111333258471 1 +340 1 5.6189938 0.9963848 0.0052250824228903484 1 +341 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +342 0 -6.470736 0.001545693 0.0022316888622042515 0 +343 0 -7.007416 0.0009043261 0.0013052570496129583 0 +344 1 8.533161 0.9998032 0.00028397130233144747 1 +345 0 -7.007416 0.0009043261 0.0013052570496129583 0 +346 0 -3.711133 0.023866279 0.034849298189560785 0 +347 0 -6.630741 0.001317447 0.0019019274473248667 0 +348 1 -0.61407375 0.3511305 1.5099208368618717 0 +349 1 2.963665 0.9509054 0.072626298277721102 1 +350 0 -4.224535 0.014421125 0.020956761314519512 0 +351 0 -5.6078997 0.0036553564 0.005283226417172892 0 +352 0 -0.5196419 0.37293598 0.67331535390783537 0 +353 1 7.1765823 0.9992363 0.0011022277999205644 1 +354 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +355 0 -4.66761 0.009307256 0.013490409215598062 0 +356 1 -1.3441267 0.20683224 2.2734669766231002 0 +357 1 11.125984 0.9999853 2.1240013733369503E-05 1 +358 1 4.329588 0.9869983 0.018880466405801065 1 +359 1 4.879279 0.9924549 0.010926571063648124 1 +360 1 13.7571335 0.9999989 1.5478446880940214E-06 1 +361 1 4.2202263 0.9855175 0.02104660253218648 1 +362 0 -4.2774982 0.013687392 0.019883118800330481 0 +363 0 -2.953164 0.04958718 0.073373796728555735 0 +364 0 -5.6078997 0.0036553564 0.005283226417172892 0 +365 0 -6.039318 0.0023775175 0.0034341166935918394 0 +366 1 12.063345 0.9999942 8.3411826797711399E-06 1 +367 1 9.634222 0.99993455 9.4421565109537245E-05 1 +368 0 -6.4123287 0.001638509 0.0023658075136661579 0 +369 0 -6.054251 0.00234236 0.00338327526506644 0 +370 0 -4.573907 0.010212206 0.014808844206685514 0 +371 0 -6.4123287 0.001638509 0.0023658075136661579 0 +372 0 -4.668415 0.009299837 0.013479605469657881 0 +373 0 -4.1111975 0.016123898 0.023451444725996713 0 +374 0 -5.2050943 0.005458573 0.0078966278438219947 0 +375 0 -7.007416 0.0009043261 0.0013052570496129583 0 +376 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +377 0 -6.6493382 0.0012932041 0.001866906609577866 0 +378 0 -4.2500477 0.014062966 0.020432582111908255 0 +379 0 -2.9373808 0.05033633 0.074511433196150584 0 +380 0 -7.007416 0.0009043261 0.0013052570496129583 0 +381 1 8.953533 0.9998707 0.00018652724247226939 1 +382 0 -4.220004 0.014485666 0.02105123932464615 0 +383 0 -6.470736 0.001545693 0.0022316888622042515 0 +384 0 -6.470736 0.001545693 0.0022316888622042515 0 +385 0 -4.255124 0.013992754 0.020329845847804801 0 +386 1 4.44446 0.9883929 0.016843462934213672 1 +387 0 -2.9901848 0.047871266 0.070771445743400579 0 +388 0 -5.81431 0.0029756578 0.0042993666724045884 0 +389 0 -3.7457647 0.023072641 0.033676803114700879 0 +390 0 -6.257917 0.0019115701 0.0027604519222911004 0 +391 1 8.564693 0.9998093 0.0002751124794602528 1 +392 0 -5.502638 0.004059459 0.0058684809920473086 0 +393 0 -7.080756 0.0008404299 0.0012129938603421102 0 +394 0 -5.9703264 0.0025469046 0.0036790937902284868 0 +395 0 -5.502638 0.004059459 0.0058684809920473086 0 +396 0 -5.07122 0.0062356335 0.009024282936362189 0 +397 0 -5.3973765 0.004508031 0.0065184177427485661 0 +398 0 -5.267047 0.0051323455 0.0074234756339299022 0 +399 0 -6.401745 0.001655914 0.0023909590824532133 0 +400 1 8.031809 0.99967515 0.0004687288599105325 1 +401 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +402 0 -3.8130236 0.021604262 0.031509975978605033 0 +403 0 -4.7785597 0.008337994 0.012079612895108123 0 +404 0 -6.135663 0.0021596157 0.0031190360999422736 0 +405 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +406 0 -4.657831 0.009397858 0.013622354083068368 0 +407 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +408 0 -4.5859423 0.0100912675 0.014632577045710669 0 +409 0 -5.2050943 0.005458573 0.0078966278438219947 0 +410 0 -6.5759974 0.0013914766 0.0020088743688413098 0 411 0 ? ? ? 0 -412 1 7.70179 0.9995482 0.00065196153701951257 1 -413 0 -3.7003446 0.02411891 0.035222726565514999 0 -414 1 5.419492 0.99559015 0.0063761374716103146 1 -415 0 -1.4183035 0.19492768 0.31280970345164211 0 -416 1 7.0347137 0.99912 0.0012701348719051887 1 -417 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -418 0 -2.3895483 0.08397317 0.12653823901896563 0 -419 0 -6.229556 0.0019664518 0.0028397832672572712 0 -420 0 -3.3565793 0.033680376 0.049427634425831944 0 -421 1 9.802544 0.9999447 7.9802157073450512E-05 1 -422 0 -3.468288 0.030228129 0.044282686546110804 0 -423 0 -3.8342748 0.021159602 0.030854450849772463 0 -424 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -425 1 12.883099 0.99999744 3.6976317320669667E-06 1 -426 0 -3.351983 0.033830285 0.049651463405332753 0 -427 1 3.4962912 0.97058207 0.043077889117461805 1 -428 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -429 0 -6.0393434 0.0023774575 0.0034340298239276387 0 -430 0 -6.3297997 0.0017792196 0.0025691575215154839 0 -431 0 -2.9899778 0.0478807 0.070785743777557433 0 -432 0 -4.204265 0.014712078 0.021382722023544836 0 -433 0 -4.9897857 0.0067610997 0.0097873289391620221 0 -434 0 3.2537518 0.9628077 4.7488528012028874 1 -435 1 5.8401175 0.99709994 0.0041899864942212704 1 -436 1 2.8955812 0.94762754 0.077607961319572816 1 -437 0 -5.3973975 0.004507937 0.0065182814231615977 0 -438 0 -4.1662054 0.01527409 0.022205875867483897 0 -439 0 -4.8607087 0.007685469 0.011130615007029858 0 -440 1 8.808516 0.9998506 0.00021559636141530833 1 -441 0 -2.4427986 0.07996677 0.12024212298095399 0 -442 0 -5.600931 0.0036808234 0.0053201028205412965 0 -443 0 -6.6222925 0.00132861 0.0019180535134382688 0 -444 0 -2.765469 0.059218936 0.088069072414956234 0 -445 0 -6.470775 0.0015456327 0.0022316017284886259 0 -446 0 -7.0074635 0.000904283 0.0013051948511978793 0 -447 0 -4.8607087 0.007685469 0.011130615007029858 0 -448 0 -7.0807962 0.00084039633 0.001212945365489941 0 -449 1 8.61463 0.9998186 0.00026169533675417532 1 -450 0 -4.403109 0.012091241 0.017550290723112521 0 -451 0 -4.8607087 0.007685469 0.011130615007029858 0 -452 0 -5.297897 0.0049772062 0.0071985199077855217 0 -453 1 7.0545425 0.9991373 0.0012451756390447772 1 -454 0 -6.3762217 0.0016986495 0.0024527169365520982 0 -455 1 0.4286108 0.6055419 0.72370134133212249 1 -456 1 9.124077 0.999891 0.00015728670782410122 1 -457 1 7.7657776 0.99957615 0.00061161394016591582 1 -458 0 -4.563208 0.010320917 0.014967307161576036 0 -459 0 -4.2657065 0.013847496 0.020117325331565697 0 -460 0 -4.224519 0.014421356 0.020957099406612927 0 -461 0 -4.3934007 0.012207759 0.017720457772906612 0 -462 0 -3.68783 0.024415229 0.035660856511145693 0 -463 0 -5.1723857 0.005639044 0.0081584455591238395 0 -464 0 -5.3973975 0.004507937 0.0065182814231615977 0 -465 1 8.705051 0.9998343 0.00023907569277659222 1 -466 1 8.096119 0.99969536 0.00043956862262736329 1 -467 1 6.187681 0.9979496 0.0029611383920793411 1 -468 0 -5.3973975 0.004507937 0.0065182814231615977 0 -469 0 -6.278531 0.0018726412 0.0027041828747360174 0 -470 0 -5.5759873 0.003773441 0.0054542217988004307 0 -471 0 -3.68783 0.024415229 0.035660856511145693 0 -472 0 -4.397353 0.0121601885 0.017650982144289801 0 -473 0 -5.3973975 0.004507937 0.0065182814231615977 0 -474 0 -4.8607087 0.007685469 0.011130615007029858 0 -475 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -476 0 -5.099897 0.0060604224 0.0087699430431780068 0 -477 0 -5.3973975 0.004507937 0.0065182814231615977 0 -478 0 -4.8500986 0.0077668102 0.011248879349031599 0 -479 1 6.4265804 0.99838465 0.0023323351816258212 1 -480 0 -4.9340415 0.0071459245 0.010346401316494924 0 -481 0 -3.5877023 0.026917238 0.039365581191928137 0 -482 1 14.304409 0.9999994 8.5991351058826534E-07 1 -483 1 9.331007 0.99991137 0.00012787476790525535 1 -484 0 -4.563208 0.010320917 0.014967307161576036 0 +412 1 7.70175 0.9995482 0.00065196153701951257 1 +413 0 -3.700317 0.024119562 0.035223690341168949 0 +414 1 5.4194355 0.9955899 0.0063764829605073153 1 +415 0 -1.4184027 0.19491212 0.31278182580827052 0 +416 1 7.034706 0.99912 0.0012701348719051887 1 +417 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +418 0 -2.389583 0.083970495 0.12653402641845476 0 +419 0 -6.2295465 0.0019664706 0.0028398105290593596 0 +420 0 -3.3565998 0.03367971 0.049426638867429623 0 +421 1 9.802483 0.9999447 7.9802157073450512E-05 1 +422 0 -3.4682703 0.030228646 0.04428345688175489 0 +423 0 -3.8341913 0.02116133 0.030856998507843483 0 +424 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +425 1 12.882942 0.99999744 3.6976317320669667E-06 1 +426 0 -3.351811 0.033835914 0.049659868584387742 0 +427 1 3.4963188 0.97058284 0.043076737348001735 1 +428 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +429 0 -6.039318 0.0023775175 0.0034341166935918394 0 +430 0 -6.3298225 0.001779179 0.0025690988018607043 0 +431 0 -2.9900103 0.047879223 0.070783502820873742 0 +432 0 -4.2042713 0.014711988 0.02138259111056056 0 +433 0 -4.989767 0.0067612245 0.0097875102090950409 0 +434 0 3.2536812 0.96280515 4.7487533854345987 1 +435 1 5.840185 0.9971001 0.0041897277699511687 1 +436 1 2.8956823 0.9476326 0.077600248117095968 1 +437 0 -5.3973765 0.004508031 0.0065184177427485661 0 +438 0 -4.1661277 0.0152752595 0.022207589624323613 0 +439 0 -4.860697 0.00768556 0.01113074702405847 0 +440 1 8.808203 0.9998505 0.00021568236559480801 1 +441 0 -2.44283 0.07996445 0.12023848951593065 0 +442 0 -5.6009226 0.003680855 0.0053201486722055345 0 +443 0 -6.62225 0.0013286663 0.0019181349102600153 0 +444 0 -2.765534 0.059215326 0.088063536759687894 0 +445 0 -6.470736 0.001545693 0.0022316888622042515 0 +446 0 -7.007416 0.0009043261 0.0013052570496129583 0 +447 0 -4.860697 0.00768556 0.01113074702405847 0 +448 0 -7.080756 0.0008404299 0.0012129938603421102 0 +449 1 8.614569 0.99981856 0.00026178134368184274 1 +450 0 -4.403137 0.012090905 0.017549799741801764 0 +451 0 -4.860697 0.00768556 0.01113074702405847 0 +452 0 -5.2978945 0.0049772183 0.0071985374621452748 0 +453 1 7.0543594 0.9991371 0.0012454338357950443 1 +454 0 -6.376228 0.001698639 0.0024527017951696023 0 +455 1 0.42855358 0.60552824 0.72373386135419882 1 +456 1 9.1238365 0.999891 0.00015728670782410122 1 +457 1 7.765707 0.99957615 0.00061161394016591582 1 +458 0 -4.5631533 0.010321477 0.01496812445290578 0 +459 0 -4.2656097 0.013848819 0.020119260056474148 0 +460 0 -4.224535 0.014421125 0.020956761314519512 0 +461 0 -4.393394 0.01220784 0.017720576112024804 0 +462 0 -3.6878552 0.024414627 0.035659966814320966 0 +463 0 -5.172369 0.0056391377 0.0081585813581601663 0 +464 0 -5.3973765 0.004508031 0.0065184177427485661 0 +465 1 8.704891 0.99983424 0.00023916169835578975 1 +466 1 8.095982 0.99969536 0.00043956862262736329 1 +467 1 6.187558 0.99794936 0.0029614830641376674 1 +468 0 -5.3973765 0.004508031 0.0065184177427485661 0 +469 0 -6.278454 0.0018727855 0.0027043915257207768 0 +470 0 -5.5759788 0.0037734734 0.005454268666255248 0 +471 0 -3.6878552 0.024414627 0.035659966814320966 0 +472 0 -4.3973575 0.012160137 0.01765090733581122 0 +473 0 -5.3973765 0.004508031 0.0065184177427485661 0 +474 0 -4.860697 0.00768556 0.01113074702405847 0 +475 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +476 0 -5.099833 0.0060608075 0.0087705020154730078 0 +477 0 -5.3973765 0.004508031 0.0065184177427485661 0 +478 0 -4.850113 0.0077667 0.011248718884428839 0 +479 1 6.4264107 0.99838436 0.0023327658339694661 1 +480 0 -4.934037 0.0071459548 0.010346445298255878 0 +481 0 -3.5877433 0.026916163 0.039363987771213436 0 +482 1 14.30418 0.9999994 8.5991351058826534E-07 1 +483 1 9.3309 0.99991137 0.00012787476790525535 1 +484 0 -4.5631533 0.010321477 0.01496812445290578 0 485 0 -5.674264 0.0034214486 0.0049445706099806612 0 -486 0 -5.5759873 0.003773441 0.0054542217988004307 0 -487 1 11.726114 0.9999919 1.1694867659388439E-05 1 -488 1 0.25589752 0.56362754 0.82718598574615088 1 -489 1 -1.4821568 0.18510188 2.4336085341210132 0 -490 0 -7.0074635 0.000904283 0.0013051948511978793 0 +486 0 -5.5759788 0.0037734734 0.005454268666255248 0 +487 1 11.725941 0.9999919 1.1694867659388439E-05 1 +488 1 0.25575066 0.5635914 0.82727844470044298 1 +489 1 -1.482214 0.18509324 2.4336758972033281 0 +490 0 -7.007416 0.0009043261 0.0013052570496129583 0 491 1 4.676321 0.9907727 0.013373942969277039 1 -492 0 -5.0392985 0.006436593 0.0093160547493154593 0 -493 1 7.964134 0.9996524 0.00050158859464882689 1 -494 0 0.41842937 0.60310733 1.3331791882434947 1 -495 0 -5.5759873 0.003773441 0.0054542217988004307 0 -496 0 -7.0807962 0.00084039633 0.001212945365489941 0 -497 0 -5.3867874 0.0045558023 0.0065876508615933917 0 -498 0 -4.9659657 0.0069229538 0.010022443686233343 0 -499 0 -4.9659657 0.0069229538 0.010022443686233343 0 -500 0 -3.4611568 0.030437874 0.044594750482780984 0 -501 0 -4.9659657 0.0069229538 0.010022443686233343 0 -502 0 -4.647871 0.009491038 0.013758065775472702 0 -503 0 -4.534534 0.0106179565 0.015400378952762167 0 -504 0 -7.0074635 0.000904283 0.0013051948511978793 0 -505 0 -5.912866 0.0026971304 0.0038963936054916032 0 -506 1 9.393385 0.99991673 0.00012013488337412622 1 -507 0 -5.8519497 0.0028660519 0.0041407754800341576 0 -508 0 -4.8607087 0.007685469 0.011130615007029858 0 -509 0 -6.470775 0.0015456327 0.0022316017284886259 0 -510 0 -7.0074635 0.000904283 0.0013051948511978793 0 -511 0 -4.429277 0.011782622 0.017099668747007298 0 -512 0 -4.8607087 0.007685469 0.011130615007029858 0 -513 0 -5.5759873 0.003773441 0.0054542217988004307 0 -514 1 9.333803 0.9999116 0.0001275307721554692 1 -515 1 6.515604 0.99852204 0.002133818139587297 1 -516 0 -7.0807962 0.00084039633 0.001212945365489941 0 -517 0 -6.6493645 0.0012931703 0.0018668576723168487 0 -518 0 -5.3184905 0.0048762504 0.0070521503051377303 0 -519 1 4.449622 0.98845196 0.016757247363327076 1 -520 0 -6.9341307 0.0009730217 0.0014044570161021121 0 -521 0 -5.3240647 0.0048492765 0.0070130450398795284 0 -522 1 4.6514864 0.9905429 0.013708652442325985 1 -523 1 6.3944283 0.99833196 0.0024084765138441796 1 -524 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -525 0 -5.6159916 0.0036260036 0.0052407246489976148 0 -526 0 -5.3973975 0.004507937 0.0065182814231615977 0 -527 0 -4.534534 0.0106179565 0.015400378952762167 0 -528 0 -3.410923 0.03195583 0.046855219598181765 0 -529 0 -5.0392985 0.006436593 0.0093160547493154593 0 -530 1 5.3436537 0.9952443 0.0068773561128845206 1 -531 0 -4.657855 0.009397636 0.013622031269102946 0 -532 0 -6.217933 0.0019893954 0.0028729494815213073 0 -533 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -534 0 -6.0393434 0.0023774575 0.0034340298239276387 0 -535 0 -5.231503 0.0053170626 0.0076913657578575002 0 -536 0 -3.9978452 0.01802431 0.026240786261966513 0 -537 0 -3.7003446 0.02411891 0.035222726565514999 0 -538 0 -4.9659657 0.0069229538 0.010022443686233343 0 -539 0 -3.8925886 0.019984946 0.029124183759263227 0 -540 0 -4.0441055 0.017223526 0.025064772138651964 0 -541 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -542 0 -4.8394885 0.007849005 0.011368394550525161 0 -543 0 -4.9659657 0.0069229538 0.010022443686233343 0 -544 0 -5.1262465 0.0059037493 0.0085425512072982886 0 -545 0 -4.429277 0.011782622 0.017099668747007298 0 -546 1 9.53874 0.999928 0.00010388126101663741 1 -547 0 -6.544108 0.0014365003 0.0020739218747718598 0 -548 0 -6.007419 0.0024543912 0.0035452905812396766 0 -549 1 5.013445 0.9933959 0.0095592667928456332 1 -550 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -551 0 -6.1446 0.0021404426 0.0030913155089767924 0 -552 0 -3.7885733 0.022127172 0.032281239435947968 0 -553 0 -2.724041 0.061569568 0.091678295805095872 0 -554 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -555 0 -2.5067987 0.07538293 0.11307210218230739 0 -556 0 -3.8764038 0.02030441 0.029594548427245702 0 -557 0 -4.224519 0.014421356 0.020957099406612927 0 -558 0 -6.0393434 0.0023774575 0.0034340298239276387 0 -559 0 -4.429277 0.011782622 0.017099668747007298 0 -560 0 -3.9978452 0.01802431 0.026240786261966513 0 -561 0 -3.9978452 0.01802431 0.026240786261966513 0 -562 0 -6.1446 0.0021404426 0.0030913155089767924 0 -563 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -564 0 -4.309522 0.013261733 0.019260634789626564 0 -565 1 10.337851 0.99996763 4.6694045347237877E-05 1 -566 0 -4.7737217 0.008378092 0.01213794979803761 0 -567 0 -4.195623 0.014837879 0.021566936543307574 0 -568 1 3.1369514 0.9583915 0.061313000673524083 1 -569 1 9.030514 0.9998803 0.00017268091542736876 1 -570 1 5.84321 0.9971089 0.0041770503375534581 1 -571 1 9.603952 0.9999325 9.7345464494158371E-05 1 -572 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -573 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -574 1 4.7993116 0.99183184 0.011832556416181512 1 -575 0 -3.7003446 0.02411891 0.035222726565514999 0 -576 0 -4.429277 0.011782622 0.017099668747007298 0 -577 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -578 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -579 0 -6.1446 0.0021404426 0.0030913155089767924 0 -580 0 -4.1317763 0.015800666 0.022977554773292386 0 -581 1 6.6461163 0.99870265 0.0018729023615442059 1 -582 1 6.0293293 0.99759865 0.0034685848342224886 1 -583 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -584 0 -3.1969008 0.039282516 0.057815851029587875 0 -585 0 -7.0074635 0.000904283 0.0013051948511978793 0 -586 1 12.05308 0.99999416 8.427174504939984E-06 1 -587 0 -4.204265 0.014712078 0.021382722023544836 0 -588 1 4.3455524 0.9872016 0.018583403862053111 1 -589 0 -4.8607087 0.007685469 0.011130615007029858 0 -590 1 2.082531 0.88919365 0.16943044261817711 1 -591 1 4.274005 0.98626536 0.019952229280350953 1 -592 1 4.843008 0.9921784 0.011328573557413945 1 -593 0 -4.563208 0.010320917 0.014967307161576036 0 -594 1 3.002367 0.95268095 0.069934961091239228 1 -595 0 -4.429277 0.011782622 0.017099668747007298 0 -596 0 -4.668465 0.009299376 0.013478934137374143 0 -597 0 -3.402843 0.032206733 0.04722919223157921 0 -598 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -599 0 -4.030929 0.017447986 0.025394312173820814 0 -600 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -601 0 -6.6493645 0.0012931703 0.0018668576723168487 0 -602 0 -4.9659657 0.0069229538 0.010022443686233343 0 -603 1 4.0770864 0.98332596 0.024258366619457607 1 -604 1 4.7631493 0.99153364 0.012266379238603733 1 -605 1 9.002562 0.9998769 0.00017758301601775289 1 -606 0 -5.079303 0.0061857454 0.0089518597426113226 0 -607 0 -7.0074635 0.000904283 0.0013051948511978793 0 -608 1 8.711237 0.9998353 0.00023761359871469259 1 -609 0 -4.8607087 0.007685469 0.011130615007029858 0 -610 1 7.8127937 0.99959564 0.00058348312768753251 1 -611 1 5.3003864 0.9950351 0.0071806598077321466 1 -612 1 15.400248 0.9999998 2.5797399935888651E-07 1 -613 0 -6.0181227 0.002428324 0.0035075914472462747 0 -614 0 -6.112676 0.0022097225 0.0031914831821281279 0 -615 0 -4.370964 0.012481298 0.018120024585944125 0 -616 0 -5.5026546 0.0040593916 0.0058683831829502422 0 +492 0 -5.039299 0.0064365896 0.009316050016199718 0 +493 1 7.9640465 0.9996524 0.00050158859464882689 1 +494 0 0.4183359 0.603085 1.3330979425018887 1 +495 0 -5.5759788 0.0037734734 0.005454268666255248 0 +496 0 -7.080756 0.0008404299 0.0012129938603421102 0 +497 0 -5.3867927 0.0045557786 0.006587616442619031 0 +498 0 -4.9659586 0.0069230027 0.010022514717742074 0 +499 0 -4.9659586 0.0069230027 0.010022514717742074 0 +500 0 -3.4611807 0.030437171 0.044593705593641481 0 +501 0 -4.9659586 0.0069230027 0.010022514717742074 0 +502 0 -4.6478777 0.009490974 0.013757973534227223 0 +503 0 -4.53454 0.010617891 0.015400283890384886 0 +504 0 -7.007416 0.0009043261 0.0013052570496129583 0 +505 0 -5.9128885 0.0026970701 0.0038963063711756288 0 +506 1 9.393265 0.99991673 0.00012013488337412622 1 +507 0 -5.851919 0.0028661392 0.0041409018059519152 0 +508 0 -4.860697 0.00768556 0.01113074702405847 0 +509 0 -6.470736 0.001545693 0.0022316888622042515 0 +510 0 -7.007416 0.0009043261 0.0013052570496129583 0 +511 0 -4.4292784 0.011782605 0.017099644273586154 0 +512 0 -4.860697 0.00768556 0.01113074702405847 0 +513 0 -5.5759788 0.0037734734 0.005454268666255248 0 +514 1 9.333614 0.9999116 0.0001275307721554692 1 +515 1 6.515238 0.9985215 0.002134593207240556 1 +516 0 -7.080756 0.0008404299 0.0012129938603421102 0 +517 0 -6.6493382 0.0012932041 0.001866906609577866 0 +518 0 -5.318432 0.0048765354 0.0070525634658966376 0 +519 1 4.4496098 0.9884518 0.016757508351226004 1 +520 0 -6.9340754 0.00097307545 0.0014045346013554871 0 +521 0 -5.3240356 0.0048494167 0.0070132482392398629 0 +522 1 4.6513977 0.99054205 0.013709867815306738 1 +523 1 6.3943024 0.9983317 0.0024088210538921761 1 +524 0 -5.502638 0.004059459 0.0058684809920473086 0 +525 0 -5.6159754 0.003626062 0.0052408092676349298 0 +526 0 -5.3973765 0.004508031 0.0065184177427485661 0 +527 0 -4.53454 0.010617891 0.015400283890384886 0 +528 0 -3.410849 0.031958118 0.046858628451947791 0 +529 0 -5.039299 0.0064365896 0.009316050016199718 0 +530 1 5.3435287 0.9952437 0.0068782201354073294 1 +531 0 -4.657831 0.009397858 0.013622354083068368 0 +532 0 -6.2179203 0.001989421 0.0028729865045729139 0 +533 0 -5.502638 0.004059459 0.0058684809920473086 0 +534 0 -6.039318 0.0023775175 0.0034341166935918394 0 +535 0 -5.2314115 0.005317547 0.0076920681723284335 0 +536 0 -3.9978604 0.01802404 0.026240389461776918 0 +537 0 -3.700317 0.024119562 0.035223690341168949 0 +538 0 -4.9659586 0.0069230027 0.010022514717742074 0 +539 0 -3.8925986 0.01998475 0.029123895846330496 0 +540 0 -4.044113 0.017223395 0.025064580736011751 0 +541 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +542 0 -4.839529 0.007848689 0.011367935461906856 0 +543 0 -4.9659586 0.0069230027 0.010022514717742074 0 +544 0 -5.12615 0.0059043146 0.0085433716250480074 0 +545 0 -4.4292784 0.011782605 0.017099644273586154 0 +546 1 9.538605 0.999928 0.00010388126101663741 1 +547 0 -6.544077 0.0014365448 0.002073986124658732 0 +548 0 -6.0073967 0.0024544462 0.0035453700495412929 0 +549 1 5.0131226 0.99339384 0.0095622965008159387 1 +550 0 -5.502638 0.004059459 0.0058684809920473086 0 +551 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +552 0 -3.7885861 0.022126894 0.032280829978766266 0 +553 0 -2.724105 0.061565876 0.09167262028876598 0 +554 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +555 0 -2.5069017 0.07537576 0.11306090709841085 0 +556 0 -3.876411 0.020304268 0.029594339965163006 0 +557 0 -4.224535 0.014421125 0.020956761314519512 0 +558 0 -6.039318 0.0023775175 0.0034341166935918394 0 +559 0 -4.4292784 0.011782605 0.017099644273586154 0 +560 0 -3.9978604 0.01802404 0.026240389461776918 0 +561 0 -3.9978604 0.01802404 0.026240389461776918 0 +562 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +563 0 -5.502638 0.004059459 0.0058684809920473086 0 +564 0 -4.3095326 0.013261596 0.019260434623768575 0 +565 1 10.337769 0.99996763 4.6694045347237877E-05 1 +566 0 -4.773676 0.008378472 0.012138502624473115 0 +567 0 -4.195627 0.014837816 0.021566845165285602 0 +568 1 3.1369228 0.95839036 0.061314705442576868 1 +569 1 9.0302925 0.9998803 0.00017268091542736876 1 +570 1 5.8429384 0.9971081 0.0041781714665405575 1 +571 1 9.603939 0.9999325 9.7345464494158371E-05 1 +572 0 -5.502638 0.004059459 0.0058684809920473086 0 +573 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +574 1 4.799197 0.99183095 0.011833856909281256 1 +575 0 -3.700317 0.024119562 0.035223690341168949 0 +576 0 -4.4292784 0.011782605 0.017099644273586154 0 +577 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +578 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +579 0 -6.1445794 0.0021404864 0.0030913787943164287 0 +580 0 -4.131735 0.015801312 0.022978502212199502 0 +581 1 6.6461525 0.99870265 0.0018729023615442059 1 +582 1 6.029065 0.99759805 0.003469446817659282 1 +583 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +584 0 -3.1969519 0.03928059 0.057812958824581623 0 +585 0 -7.007416 0.0009043261 0.0013052570496129583 0 +586 1 12.052902 0.99999416 8.427174504939984E-06 1 +587 0 -4.2042713 0.014711988 0.02138259111056056 0 +588 1 4.3455086 0.98720104 0.018584187817591195 1 +589 0 -4.860697 0.00768556 0.01113074702405847 0 +590 1 2.0824509 0.8891858 0.16944320800974624 1 +591 1 4.273983 0.98626506 0.019952665224579537 1 +592 1 4.842909 0.9921776 0.011329700257702026 1 +593 0 -4.5631533 0.010321477 0.01496812445290578 0 +594 1 3.0022697 0.95267653 0.069941640528718624 1 +595 0 -4.4292784 0.011782605 0.017099644273586154 0 +596 0 -4.668415 0.009299837 0.013479605469657881 0 +597 0 -3.4027734 0.0322089 0.047232424262708413 0 +598 0 -5.502638 0.004059459 0.0058684809920473086 0 +599 0 -4.0308228 0.01744981 0.025396989690649865 0 +600 0 -5.502638 0.004059459 0.0058684809920473086 0 +601 0 -6.6493382 0.0012932041 0.001866906609577866 0 +602 0 -4.9659586 0.0069230027 0.010022514717742074 0 +603 1 4.077015 0.98332477 0.024260115609745794 1 +604 1 4.7629147 0.99153167 0.012269241185438283 1 +605 1 9.002628 0.9998769 0.00017758301601775289 1 +606 0 -5.0792956 0.0061857896 0.0089519239615415361 0 +607 0 -7.007416 0.0009043261 0.0013052570496129583 0 +608 1 8.711155 0.9998353 0.00023761359871469259 1 +609 0 -4.860697 0.00768556 0.01113074702405847 0 +610 1 7.812504 0.9995955 0.00058365517991933137 1 +611 1 5.3003216 0.9950348 0.0071810919097619157 1 +612 1 15.399965 0.9999998 2.5797399935888651E-07 1 +613 0 -6.0181503 0.0024282571 0.0035074948082411965 0 +614 0 -6.1126585 0.0022097614 0.003191539402263821 0 +615 0 -4.3708715 0.012482438 0.018121689956984946 0 +616 0 -5.502638 0.004059459 0.0058684809920473086 0 617 0 ? ? ? 0 -618 0 -4.9659657 0.0069229538 0.010022443686233343 0 -619 0 -4.429277 0.011782622 0.017099668747007298 0 -620 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -621 0 -0.45594597 0.387948 0.70827388031103966 0 -622 0 -2.4222088 0.08149477 0.12264016025567508 0 -623 0 -7.0074635 0.000904283 0.0013051948511978793 0 -624 0 -4.184515 0.015001129 0.021806024451924741 0 -625 0 -4.110237 0.01613914 0.023473794861896838 0 -626 1 4.3905373 0.9877577 0.017770932328023542 1 -627 0 -4.9300895 0.0071740183 0.010387224410673452 0 -628 0 -6.470775 0.0015456327 0.0022316017284886259 0 -629 0 -5.3973975 0.004507937 0.0065182814231615977 0 -630 0 -3.718408 0.02369738 0.034599692246208076 0 -631 0 -4.429277 0.011782622 0.017099668747007298 0 -632 0 -7.0074635 0.000904283 0.0013051948511978793 0 -633 1 3.400981 0.9677352 0.047315798729693895 1 -634 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -635 0 -5.0415835 0.0064219963 0.0092948600126800188 0 -636 1 8.94837 0.99987006 0.0001874732696668834 1 -637 0 -3.0414915 0.045586236 0.067313245725512838 0 -638 0 -5.3973975 0.004507937 0.0065182814231615977 0 -639 0 -4.224519 0.014421356 0.020957099406612927 0 -640 0 -4.761208 0.0084826965 0.012290144952248007 0 -641 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -642 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -643 0 -7.0074635 0.000904283 0.0013051948511978793 0 -644 0 -6.470775 0.0015456327 0.0022316017284886259 0 -645 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -646 0 -6.6993523 0.0012301942 0.0017758876872952787 0 -647 0 -6.371274 0.0017070603 0.0024648718167361995 0 -648 1 11.00214 0.9999833 2.4077772041939022E-05 1 -649 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -650 0 -4.387144 0.012283436 0.017830990738932077 0 -651 0 -5.9292316 0.0026534658 0.0038332299316791718 0 -652 0 -4.204265 0.014712078 0.021382722023544836 0 -653 0 -4.9659657 0.0069229538 0.010022443686233343 0 -654 0 -5.071223 0.006235616 0.009024257247501357 0 -655 0 -5.5026546 0.0040593916 0.0058683831829502422 0 -656 0 -4.429277 0.011782622 0.017099668747007298 0 -657 0 -1.2109537 0.22953235 0.37619370476645442 0 -658 1 7.375309 0.99937385 0.00090362180313618902 1 -659 0 -7.0074635 0.000904283 0.0013051948511978793 0 -660 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -661 0 -4.534534 0.0106179565 0.015400378952762167 0 -662 0 -5.9810295 0.0025198585 0.0036399754528292716 0 -663 0 -5.9810295 0.0025198585 0.0036399754528292716 0 -664 0 -4.866465 0.0076416926 0.011066971342776052 0 -665 0 -7.0074635 0.000904283 0.0013051948511978793 0 -666 0 -3.862595 0.020580921 0.030001795465253788 0 -667 0 -5.071223 0.006235616 0.009024257247501357 0 -668 1 1.7462997 0.8514855 0.23194614810831629 1 -669 1 6.513338 0.9985187 0.0021386407895282642 1 -670 1 4.3529015 0.9872941 0.018448221455131173 1 -671 0 -4.742518 0.008641346 0.012521004960698167 0 -672 0 -5.6079116 0.0036553128 0.0052831633727484281 0 -673 0 -4.4686546 0.011332822 0.016443156821417514 0 -674 0 -6.5760317 0.0013914289 0.0020088054126509622 0 -675 0 -4.5048947 0.010933883 0.015861128888792996 0 -676 0 -6.278531 0.0018726412 0.0027041828747360174 0 -677 0 -4.8607087 0.007685469 0.011130615007029858 0 -678 0 -7.0074635 0.000904283 0.0013051948511978793 0 -679 0 -6.470775 0.0015456327 0.0022316017284886259 0 -680 1 15.284569 0.99999976 3.4396534272948301E-07 1 -681 1 7.982956 0.9996589 0.00049221231130512611 1 -682 0 -3.7728333 0.02247032 0.032787587809673793 0 -683 0 -7.0074635 0.000904283 0.0013051948511978793 0 -684 0 -7.0074635 0.000904283 0.0013051948511978793 0 -685 0 -7.0074635 0.000904283 0.0013051948511978793 0 -686 0 -7.0074635 0.000904283 0.0013051948511978793 0 -687 0 -5.052631 0.0063518886 0.009193065922136142 0 -688 0 -5.3973975 0.004507937 0.0065182814231615977 0 -689 0 -4.404111 0.01207928 0.017532823587725396 0 -690 0 -6.371274 0.0017070603 0.0024648718167361995 0 -691 1 2.8413534 0.94487 0.081812252645891606 1 -692 0 -5.9340863 0.0026406492 0.0038146903700998738 0 -693 0 -5.130747 0.005877395 0.0085043049849836787 0 -694 0 -5.475582 0.0041703256 0.0060290884270468486 0 -695 0 -6.470775 0.0015456327 0.0022316017284886259 0 -696 1 5.1025257 0.9939554 0.0087470152271597566 1 -697 1 3.3554554 0.966283 0.049482279956140095 1 -698 1 4.277033 0.9863063 0.019892331795815305 1 +618 0 -4.9659586 0.0069230027 0.010022514717742074 0 +619 0 -4.4292784 0.011782605 0.017099644273586154 0 +620 0 -5.502638 0.004059459 0.0058684809920473086 0 +621 0 -0.4560032 0.38793442 0.70824184740412566 0 +622 0 -2.4222136 0.08149441 0.12263959853012815 0 +623 0 -7.007416 0.0009043261 0.0013052570496129583 0 +624 0 -4.1845384 0.015000784 0.021805518379387972 0 +625 0 -4.1101313 0.01614082 0.023476258505542608 0 +626 1 4.390461 0.9877567 0.017772325242394194 1 +627 0 -4.9300737 0.0071741305 0.010387387486131338 0 +628 0 -6.470736 0.001545693 0.0022316888622042515 0 +629 0 -5.3973765 0.004508031 0.0065184177427485661 0 +630 0 -3.7183456 0.023698825 0.034601828152783121 0 +631 0 -4.4292784 0.011782605 0.017099644273586154 0 +632 0 -7.007416 0.0009043261 0.0013052570496129583 0 +633 1 3.4008121 0.9677299 0.047323618283400176 1 +634 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +635 0 -5.041425 0.0064230068 0.0092963272577526873 0 +636 1 8.948109 0.99987006 0.0001874732696668834 1 +637 0 -3.0414162 0.04558951 0.067318195524774263 0 +638 0 -5.3973765 0.004508031 0.0065184177427485661 0 +639 0 -4.224535 0.014421125 0.020956761314519512 0 +640 0 -4.761215 0.008482636 0.012290056870136968 0 +641 0 -5.502638 0.004059459 0.0058684809920473086 0 +642 0 -5.502638 0.004059459 0.0058684809920473086 0 +643 0 -7.007416 0.0009043261 0.0013052570496129583 0 +644 0 -6.470736 0.001545693 0.0022316888622042515 0 +645 0 -5.502638 0.004059459 0.0058684809920473086 0 +646 0 -6.6992884 0.0012302727 0.0017760010262468937 0 +647 0 -6.371254 0.0017070944 0.0024649211107645097 0 +648 1 11.001965 0.9999833 2.4077772041939022E-05 1 +649 0 -5.502638 0.004059459 0.0058684809920473086 0 +650 0 -4.387138 0.012283511 0.017831100925173125 0 +651 0 -5.9292526 0.0026534102 0.0038331494371278917 0 +652 0 -4.2042713 0.014711988 0.02138259111056056 0 +653 0 -4.9659586 0.0069230027 0.010022514717742074 0 +654 0 -5.07122 0.0062356335 0.009024282936362189 0 +655 0 -5.502638 0.004059459 0.0058684809920473086 0 +656 0 -4.4292784 0.011782605 0.017099644273586154 0 +657 0 -1.2108526 0.22955023 0.37622718793328674 0 +658 1 7.3751793 0.9993738 0.00090370784834111395 1 +659 0 -7.007416 0.0009043261 0.0013052570496129583 0 +660 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +661 0 -4.53454 0.010617891 0.015400283890384886 0 +662 0 -5.9809103 0.0025201582 0.0036404088529536712 0 +663 0 -5.9809103 0.0025201582 0.0036404088529536712 0 +664 0 -4.866476 0.0076416098 0.011066850840250858 0 +665 0 -7.007416 0.0009043261 0.0013052570496129583 0 +666 0 -3.8625927 0.02058097 0.03000186680137043 0 +667 0 -5.07122 0.0062356335 0.009024282936362189 0 +668 1 1.7461681 0.85146886 0.23197432452980324 1 +669 1 6.513275 0.9985186 0.0021388130273243158 1 +670 1 4.3527374 0.98729205 0.018451182789682705 1 +671 0 -4.742555 0.008641028 0.012520541439176703 0 +672 0 -5.6078997 0.0036553564 0.005283226417172892 0 +673 0 -4.468645 0.011332929 0.016443313108259736 0 +674 0 -6.5759974 0.0013914766 0.0020088743688413098 0 +675 0 -4.5047455 0.010935497 0.015863483115388894 0 +676 0 -6.278454 0.0018727855 0.0027043915257207768 0 +677 0 -4.860697 0.00768556 0.01113074702405847 0 +678 0 -7.007416 0.0009043261 0.0013052570496129583 0 +679 0 -6.470736 0.001545693 0.0022316888622042515 0 +680 1 15.284296 0.99999976 3.4396534272948301E-07 1 +681 1 7.982851 0.9996588 0.00049229833197626739 1 +682 0 -3.772853 0.02246989 0.032786952790854099 0 +683 0 -7.007416 0.0009043261 0.0013052570496129583 0 +684 0 -7.007416 0.0009043261 0.0013052570496129583 0 +685 0 -7.007416 0.0009043261 0.0013052570496129583 0 +686 0 -7.007416 0.0009043261 0.0013052570496129583 0 +687 0 -5.052623 0.00635194 0.0091931402933308863 0 +688 0 -5.3973765 0.004508031 0.0065184177427485661 0 +689 0 -4.404144 0.012078887 0.017532249649779036 0 +690 0 -6.371254 0.0017070944 0.0024649211107645097 0 +691 1 2.8413935 0.9448721 0.081809067347296954 1 +692 0 -5.9340563 0.0026407284 0.0038148048797132152 0 +693 0 -5.130742 0.005877423 0.0085043455317272768 0 +694 0 -5.4755497 0.00417046 0.006029283392418496 0 +695 0 -6.470736 0.001545693 0.0022316888622042515 0 +696 1 5.102537 0.99395543 0.0087469287128904821 1 +697 1 3.3553085 0.96627825 0.049489399322663288 1 +698 1 4.276967 0.9863054 0.019893639574599043 1 diff --git a/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris-out.txt b/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris-out.txt new file mode 100644 index 0000000000..10721bd5ce --- /dev/null +++ b/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris-out.txt @@ -0,0 +1,68 @@ +maml.exe TrainTest test=%Data% tr=MulticlassLogisticRegression{l1=0.001 l2=0.1 ot=1e-3 nt=1 nn=+} norm=No dout=%Output% data=%Data% out=%Output% seed=1 xf=Term{col=Label} +Not adding a normalizer. +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 12 of 15 weights. +Not training a calibrator because it is not needed. + +Confusion table + ||======================== +PREDICTED || 0 | 1 | 2 | Recall +TRUTH ||======================== + 0 || 50 | 0 | 0 | 1.0000 + 1 || 0 | 48 | 2 | 0.9600 + 2 || 0 | 2 | 48 | 0.9600 + ||======================== +Precision ||1.0000 |0.9600 |0.9600 | +Accuracy(micro-avg): 0.973333 +Accuracy(macro-avg): 0.973333 +Log-loss: 0.099452 +Log-loss reduction: 0.909475 + +OVERALL RESULTS +--------------------------------------- +Accuracy(micro-avg): 0.973333 (0.0000) +Accuracy(macro-avg): 0.973333 (0.0000) +Log-loss: 0.099452 (0.0000) +Log-loss reduction: 0.909475 (0.0000) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + +--- Progress log --- +[1] 'Building term dictionary' started. +[1] (%Time%) 150 examples Total Terms: 3 +[1] 'Building term dictionary' finished in %Time%. +[2] 'LBFGS data prep' started. +[2] 'LBFGS data prep' finished in %Time%. +[3] 'LBFGS Optimizer' started. +[3] (%Time%) 0 iterations Loss: 1.0986123085021973 +[3] (%Time%) 1 iterations Loss: 1.0638996362686157 Improvement: 0.03471 +[3] (%Time%) 2 iterations Loss: 1.0165412425994873 Improvement: 0.04483 +[3] (%Time%) 3 iterations Loss: 0.9443140625953674 Improvement: 0.0657 +[3] (%Time%) 4 iterations Loss: 0.6682096123695374 Improvement: 0.2241 +[3] (%Time%) 5 iterations Loss: 0.553279459476471 Improvement: 0.1421 +[3] (%Time%) 6 iterations Loss: 0.42720964550971985 Improvement: 0.1301 +[3] (%Time%) 7 iterations Loss: 0.3354315459728241 Improvement: 0.1014 +[3] (%Time%) 8 iterations Loss: 0.2713882327079773 Improvement: 0.07337 +[3] (%Time%) 9 iterations Loss: 0.21875490248203278 Improvement: 0.05782 +[3] (%Time%) 10 iterations Loss: 0.19283106923103333 Improvement: 0.0339 +[3] (%Time%) 11 iterations Loss: 0.1848226934671402 Improvement: 0.01448 +[3] (%Time%) 12 iterations Loss: 0.18257759511470795 Improvement: 0.005304 +[3] (%Time%) 13 iterations Loss: 0.18094150722026825 Improvement: 0.002553 +[3] (%Time%) 14 iterations Loss: 0.17891088128089905 Improvement: 0.002161 +[3] (%Time%) 15 iterations Loss: 0.17133614420890808 Improvement: 0.006221 +[3] (%Time%) 16 iterations Loss: 0.1575724482536316 Improvement: 0.01188 +[3] (%Time%) 17 iterations Loss: 0.15349344909191132 Improvement: 0.006029 +[3] (%Time%) 18 iterations Loss: 0.15131297707557678 Improvement: 0.003143 +[3] (%Time%) 19 iterations Loss: 0.14684413373470306 Improvement: 0.004137 +[3] (%Time%) 20 iterations Loss: 0.14376644790172577 Improvement: 0.003343 +[3] (%Time%) 21 iterations Loss: 0.14153197407722473 Improvement: 0.002512 +[3] (%Time%) 22 iterations Loss: 0.1407090276479721 Improvement: 0.001245 +[3] (%Time%) 23 iterations Loss: 0.14069536328315735 Improvement: 0.0003215 +[3] 'LBFGS Optimizer' finished in %Time%. +[4] 'Saving model' started. +[4] 'Saving model' finished in %Time%. diff --git a/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris-rp.txt b/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris-rp.txt new file mode 100644 index 0000000000..c07eb23a28 --- /dev/null +++ b/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris-rp.txt @@ -0,0 +1,4 @@ +MulticlassLogisticRegression +Accuracy(micro-avg) Accuracy(macro-avg) Log-loss Log-loss reduction /l2 /l1 /ot /nt /nn Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.973333 0.973333 0.099452 0.909475 0.1 0.001 0.001 1 + MulticlassLogisticRegression %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=MulticlassLogisticRegression{l1=0.001 l2=0.1 ot=1e-3 nt=1 nn=+} norm=No dout=%Output% data=%Data% out=%Output% seed=1 xf=Term{col=Label} /l2:0.1;/l1:0.001;/ot:0.001;/nt:1;/nn:+ + diff --git a/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris.txt b/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris.txt index ebaccc4d62..d44cac07f2 100644 --- a/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris.txt +++ b/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/LogisticRegression-Non-Negative-TrainTest-iris.txt @@ -1,151 +1,151 @@ Instance Label Assigned Log-loss #1 Score #2 Score #3 Score #1 Class #2 Class #3 Class -0 0 0 0.0055160384679411277 0.99449915 0.005502038 5.1773974E-11 0 1 2 -1 0 0 0.023864730958411882 0.9764178 0.023581393 1.0470911E-09 0 1 2 -2 0 0 0.0093345440359657913 0.9907089 0.009291213 3.3886932E-10 0 1 2 -3 0 0 0.023485169105185572 0.97678846 0.023210973 2.9965708E-09 0 1 2 -4 0 0 0.0040397473706441957 0.9959684 0.0040320004 4.1783424E-11 0 1 2 -5 0 0 0.0045280297264245848 0.9954822 0.004518237 6.792683E-11 0 1 2 -6 0 0 0.0069865892171306161 0.99303776 0.006961399 5.4606175E-10 0 1 2 -7 0 0 0.010116555796003303 0.98993444 0.010064013 2.2703014E-10 0 1 2 -8 0 0 0.030149466155855287 0.9703005 0.029698554 6.8450254E-09 0 1 2 -9 0 0 0.024381664764845521 0.97591317 0.02408588 8.098426E-10 0 1 2 -10 0 0 0.0043411169825802809 0.9956683 0.0043306584 1.7070477E-11 0 1 2 -11 0 0 0.013576521322325076 0.9865152 0.013485671 8.031634E-10 0 1 2 -12 0 0 0.023572921206410413 0.97670275 0.023298172 8.6137114E-10 0 1 2 -13 0 0 0.0085086580176359974 0.99152744 0.008472632 4.1598563E-10 0 1 2 -14 0 0 0.0007095469417312554 0.9992907 0.00070933235 1.5176845E-13 0 1 2 -15 0 0 0.00055310522407816078 0.99944705 0.00055347907 6.048726E-13 0 1 2 -16 0 0 0.0012550477831004697 0.99874574 0.0012540518 3.9664222E-12 0 1 2 -17 0 0 0.0055160384679411277 0.99449915 0.0055009676 8.31305E-11 0 1 2 -18 0 0 0.0063400554390819954 0.99368 0.0063216076 3.0053654E-11 0 1 2 -19 0 0 0.0030975582921822227 0.99690723 0.0030934948 3.8560665E-11 0 1 2 -20 0 0 0.020122528904036908 0.9800786 0.019920269 3.064467E-10 0 1 2 -21 0 0 0.0041771031887722987 0.9958316 0.0041694827 1.0134493E-10 0 1 2 -22 0 0 0.0010643302939017342 0.99893624 0.0010639144 7.4156245E-12 0 1 2 -23 0 0 0.02605629788130279 0.97428024 0.025720494 4.7576445E-09 0 1 2 -24 0 0 0.035209627128022683 0.965403 0.03459665 6.6329586E-09 0 1 2 -25 0 0 0.045448274212285186 0.955569 0.04443145 3.216728E-09 0 1 2 -26 0 0 0.013916015058125409 0.98618037 0.013818858 1.1871322E-09 0 1 2 -27 0 0 0.0076942651707202743 0.99233526 0.0076657226 7.9652826E-11 0 1 2 -28 0 0 0.007534024516302436 0.9924943 0.0075039454 6.411831E-11 0 1 2 -29 0 0 0.024288162137741393 0.9760044 0.02399573 2.8173113E-09 0 1 2 -30 0 0 0.033046722035427502 0.96749336 0.03250687 3.46561E-09 0 1 2 -31 0 0 0.010627754848336371 0.9894285 0.010573362 1.9241796E-10 0 1 2 -32 0 0 0.0012779052922244011 0.9987229 0.0012773476 2.5791548E-12 0 1 2 -33 0 0 0.00071336435395046255 0.9992869 0.0007130818 5.394398E-13 0 1 2 -34 0 0 0.024381664764845521 0.97591317 0.02408588 8.098426E-10 0 1 2 -35 0 0 0.0070380898792076199 0.9929866 0.007013349 7.24567E-11 0 1 2 -36 0 0 0.0042114001992704566 0.99579746 0.004201503 8.375258E-12 0 1 2 -37 0 0 0.024381664764845521 0.97591317 0.02408588 8.098426E-10 0 1 2 -38 0 0 0.016315454412074732 0.9838169 0.01618225 2.0803645E-09 0 1 2 -39 0 0 0.010242464293569139 0.9898098 0.010190028 1.7190988E-10 0 1 2 -40 0 0 0.0039520171010716818 0.9960558 0.003945863 5.4001064E-11 0 1 2 -41 0 0 0.12730980592945779 0.88046086 0.119539626 7.181209E-08 0 1 2 -42 0 0 0.0089912491214826969 0.99104905 0.0089505045 7.80488E-10 0 1 2 -43 0 0 0.010326412100003242 0.9897267 0.010272516 1.874473E-09 0 1 2 -44 0 0 0.011150365797478341 0.98891157 0.011090281 1.0552612E-09 0 1 2 -45 0 0 0.02356529293651816 0.9767102 0.023289332 2.2207163E-09 0 1 2 -46 0 0 0.0042686243571466928 0.9957405 0.0042613307 4.8837566E-11 0 1 2 -47 0 0 0.012689595644547294 0.9873906 0.012608856 9.079638E-10 0 1 2 -48 0 0 0.0042877197577386979 0.99572146 0.0042767935 2.2542229E-11 0 1 2 -49 0 0 0.0099010247223913338 0.9901478 0.009852205 1.827634E-10 0 1 2 -50 1 1 0.0090408081662831513 0.99099994 0.007567417 0.0014314867 1 2 0 -51 1 1 0.034511583658046456 0.96607715 0.031063149 0.002861517 1 2 0 -52 1 1 0.04304121150554157 0.9578719 0.04158025 0.00054618425 1 2 0 -53 1 1 0.13135529742102009 0.87690616 0.122118875 0.0009759098 1 2 0 -54 1 1 0.074932084511623434 0.9278065 0.07160018 0.00059335935 1 2 0 -55 1 1 0.1890525667021884 0.827743 0.1714483 0.00080681086 1 2 0 -56 1 1 0.11877056857344356 0.8880115 0.1100965 0.0018910082 1 2 0 -57 1 1 0.025358193239084965 0.9749606 0.01495395 0.010085882 1 0 2 -58 1 1 0.019096356683429662 0.9810848 0.01808086 0.0008358488 1 2 0 -59 1 1 0.15862658959532094 0.85331494 0.14217468 0.0045111584 1 2 0 -60 1 1 0.038318653357595697 0.9624062 0.035282373 0.0023114693 1 2 0 -61 1 1 0.065311423102221738 0.9367757 0.058971565 0.0042529916 1 2 0 -62 1 1 0.010269141347627856 0.9897834 0.009452066 0.0007660691 1 2 0 -63 1 1 0.17183304995881135 0.84211975 0.1573262 0.00055418257 1 2 0 -64 1 1 0.030368791940673646 0.9700877 0.023280893 0.0066331304 1 0 2 -65 1 1 0.00975415306436043 0.99029326 0.0068194265 0.0028855118 1 2 0 -66 1 1 0.39629741859747109 0.67280656 0.32598528 0.0012099475 1 2 0 -67 1 1 0.012058247747154605 0.98801416 0.009438771 0.0025457656 1 2 0 -68 1 1 0.3368797650804119 0.7139947 0.28589886 0.00010822646 1 2 0 -69 1 1 0.021083825783822494 0.9791369 0.018163126 0.0027012476 1 2 0 -70 1 2 1.0416832150942417 0.6467134 0.35286024 0.00042478985 2 1 0 -71 1 1 0.013755801712349471 0.9863384 0.00910499 0.0045568454 1 2 0 -72 1 1 0.58514020696109537 0.55702776 0.44291526 5.669436E-05 1 2 0 -73 1 1 0.084747339774486161 0.9187444 0.080806226 0.00044783644 1 2 0 -74 1 1 0.012447376934475682 0.9876298 0.010109726 0.0022619804 1 2 0 -75 1 1 0.013278939705380893 0.98680884 0.011033021 0.0021575652 1 2 0 -76 1 1 0.04319766018798589 0.95772207 0.041966196 0.0003101741 1 2 0 -77 1 1 0.30690002516055565 0.73572415 0.2640443 0.00023127515 1 2 0 -78 1 1 0.16997907799864528 0.84368247 0.1552497 0.0010691443 1 2 0 -79 1 1 0.014675126081593009 0.985432 0.013093817 0.0014749394 1 0 2 -80 1 1 0.022987361370576598 0.97727484 0.01993399 0.00278951 1 2 0 -81 1 1 0.012458783376888797 0.9876185 0.008495432 0.003886289 1 2 0 -82 1 1 0.016071326268181901 0.9840571 0.011121038 0.004822609 1 2 0 -83 1 2 1.707477571538714 0.8186583 0.18132259 1.8361572E-05 2 1 0 -84 1 1 0.62495808673272679 0.53528386 0.4637305 0.0009871349 1 2 0 -85 1 1 0.11122891687722759 0.8947339 0.1001901 0.0050765327 1 2 0 -86 1 1 0.036052681021758022 0.9645895 0.03433999 0.0010723303 1 2 0 -87 1 1 0.063015011349415673 0.9389294 0.060807556 0.00026142705 1 2 0 -88 1 1 0.044994341810109487 0.9560029 0.037784368 0.0062121437 1 2 0 -89 1 1 0.092109667787323657 0.9120051 0.086147815 0.0018485248 1 2 0 -90 1 1 0.20783613448356417 0.81234014 0.18704556 0.0006146012 1 2 0 -91 1 1 0.10023498504355982 0.9046248 0.09426649 0.0011077761 1 2 0 -92 1 1 0.022636421664178376 0.97761786 0.01980763 0.0025746985 1 2 0 -93 1 1 0.020406702516074512 0.9798001 0.0109967375 0.009203869 1 0 2 -94 1 1 0.10109711620080604 0.90384525 0.09442052 0.0017330045 1 2 0 -95 1 1 0.031333905261789932 0.9691519 0.02633722 0.0045093265 1 2 0 -96 1 1 0.054904961042088807 0.9465751 0.050160136 0.0032641925 1 2 0 -97 1 1 0.020442959924209147 0.9797646 0.017932491 0.0023010836 1 2 0 -98 1 1 0.053829215135703323 0.9475939 0.05015344 0.00225256 1 0 2 -99 1 1 0.046024295999239594 0.95501876 0.04161595 0.0033648298 1 2 0 -100 2 2 0.00071718178074236128 0.9992831 0.0007153486 2.3431886E-08 2 1 0 -101 2 2 0.029472378645200227 0.9709577 0.029038547 3.0171875E-06 2 1 0 -102 2 2 0.0393600403713796 0.9614045 0.038595207 6.408885E-07 2 1 0 -103 2 2 0.042228681999026033 0.9586505 0.04134742 1.4739417E-06 2 1 0 -104 2 2 0.0064353741478142112 0.9935853 0.006415478 1.5842724E-07 2 1 0 -105 2 2 0.011154163002271514 0.9889078 0.011091466 1.8252322E-08 2 1 0 -106 2 2 0.038921380516359176 0.9618263 0.038155746 1.6744794E-05 2 1 0 -107 2 2 0.050231903865451351 0.95100886 0.04899014 1.6252797E-07 2 1 0 -108 2 2 0.028682695065745022 0.97172475 0.028276091 1.5199748E-07 2 1 0 -109 2 2 0.011791212573983524 0.98827803 0.011722299 6.1125087E-07 2 1 0 -110 2 2 0.31905363806129555 0.72683656 0.27304813 0.000116325275 2 1 0 -111 2 2 0.07549667881010734 0.9272828 0.072712734 3.6851195E-06 2 1 0 -112 2 2 0.076789840257624864 0.92608446 0.07391078 4.6068785E-06 2 1 0 -113 2 2 0.013851165145911459 0.9862443 0.013756296 1.0960175E-06 2 1 0 -114 2 2 0.0033950556005307988 0.9966107 0.0033878458 4.755091E-07 2 1 0 -115 2 2 0.03064347778045929 0.9698213 0.030172396 6.850469E-06 2 1 0 -116 2 2 0.12950899319139736 0.8785267 0.121466115 7.857308E-06 2 1 0 -117 2 2 0.029441808148158395 0.9709874 0.029012194 3.762749E-07 2 1 0 -118 2 2 0.00083159205288146009 0.99916875 0.00083038886 1.5524602E-10 2 1 0 -119 2 2 0.18149944608911653 0.8340187 0.16597763 5.176206E-06 2 1 0 -120 2 2 0.027599360244525448 0.972778 0.027219418 1.605549E-06 2 1 0 -121 2 2 0.027206434952982018 0.9731603 0.026830971 7.3364695E-06 2 1 0 -122 2 2 0.011005419791721024 0.9890549 0.010943153 7.0800863E-09 2 1 0 -123 2 2 0.36999896999840359 0.69073504 0.30920705 5.735062E-05 2 1 0 -124 2 2 0.047847764215344722 0.9532789 0.046716593 3.8119588E-06 2 1 0 -125 2 2 0.20049289074676657 0.8183273 0.18166876 3.932513E-06 2 1 0 -126 2 2 0.47024913514888667 0.6248466 0.3750236 0.00013107271 2 1 0 -127 2 2 0.37044139869226145 0.6904295 0.30942416 0.00014464991 2 1 0 -128 2 2 0.011402157721508203 0.9886626 0.011337829 2.959167E-07 2 1 0 -129 2 2 0.61341094273731411 0.5415007 0.45849052 1.03566035E-05 2 1 0 -130 2 2 0.074195879113551133 0.9284898 0.07150847 3.3010758E-07 2 1 0 -131 2 2 0.36711882425272957 0.6927273 0.3072634 1.0183718E-05 2 1 0 -132 2 2 0.0071143853619847876 0.99291086 0.007090189 1.8508953E-07 2 1 0 -133 2 1 0.70938490731195092 0.507985 0.4919467 6.684E-05 1 2 0 -134 2 2 0.085842140845050208 0.9177391 0.082260855 1.222486E-06 2 1 0 -135 2 2 0.040000927528960964 0.96078855 0.039211515 3.1772186E-07 2 1 0 -136 2 2 0.0066146987210017709 0.99340713 0.0065912344 1.0530171E-06 2 1 0 -137 2 2 0.11827848787180953 0.8884486 0.11154141 9.8605E-06 2 1 0 -138 2 2 0.40225979543659157 0.66880697 0.3309761 0.00021604139 2 1 0 -139 2 2 0.17508696641630789 0.839384 0.1605988 1.8394523E-05 2 1 0 -140 2 2 0.011779813743593867 0.9882893 0.011711036 7.23893E-07 2 1 0 -141 2 2 0.21421052275020513 0.80717844 0.19276378 5.790216E-05 2 1 0 -142 2 2 0.029472378645200227 0.9709577 0.029038547 3.0171875E-06 2 1 0 -143 2 2 0.0095519990430713575 0.9904935 0.009506778 2.986748E-07 2 1 0 -144 2 2 0.0073394639370607329 0.9926874 0.0073135537 5.9723027E-07 2 1 0 -145 2 2 0.071865099775285421 0.93065643 0.06933218 1.1476711E-05 2 1 0 -146 2 2 0.12049862656713971 0.8864783 0.11351164 8.385402E-06 2 1 0 -147 2 2 0.15922545922324882 0.85280406 0.14717093 2.4967045E-05 2 1 0 -148 2 2 0.017231011314152864 0.9829166 0.017076416 5.251435E-06 2 1 0 -149 2 2 0.10884858930380203 0.8968662 0.103108585 2.5997891E-05 2 1 0 +0 0 0 0.0061950852433387601 0.99382406 0.0061750147 6.415856E-11 0 1 2 +1 0 0 0.024190882270233992 0.9760994 0.02390031 1.1836706E-09 0 1 2 +2 0 0 0.010334000269108155 0.9897192 0.0102813365 4.0845438E-10 0 1 2 +3 0 0 0.024387344827267593 0.9759076 0.024092909 3.3243117E-09 0 1 2 +4 0 0 0.0046844358446896913 0.9953265 0.004672967 5.2640076E-11 0 1 2 +5 0 0 0.0050735213310076762 0.9949393 0.0050600115 8.021832E-11 0 1 2 +6 0 0 0.0079841810154562595 0.9920476 0.007952063 6.505245E-10 0 1 2 +7 0 0 0.010911954320681469 0.98914737 0.010853416 2.6727126E-10 0 1 2 +8 0 0 0.031099303189334446 0.9693793 0.030621191 7.542666E-09 0 1 2 +9 0 0 0.024698270577283184 0.97560424 0.024394816 9.156438E-10 0 1 2 +10 0 0 0.0048637461544768163 0.99514806 0.0048530777 2.1329287E-11 0 1 2 +11 0 0 0.014535897462350887 0.98556924 0.014430009 9.1312186E-10 0 1 2 +12 0 0 0.024089826265429194 0.976198 0.023801321 9.860573E-10 0 1 2 +13 0 0 0.0097446432661685064 0.9903027 0.009698148 5.199548E-10 0 1 2 +14 0 0 0.00088880209796682742 0.9991116 0.0008885167 2.214887E-13 0 1 2 +15 0 0 0.00071718178074236128 0.9992831 0.00071511255 8.3642067E-13 0 1 2 +16 0 0 0.0015487737280443089 0.9984524 0.0015481457 5.322795E-12 0 1 2 +17 0 0 0.0061950852433387601 0.99382406 0.0061750147 1.0156496E-10 0 1 2 +18 0 0 0.0067634503253886933 0.9932594 0.006742421 3.5515146E-11 0 1 2 +19 0 0 0.0036468762471827455 0.99635977 0.0036403935 4.8154106E-11 0 1 2 +20 0 0 0.020011910373325573 0.980187 0.019812541 3.4069497E-10 0 1 2 +21 0 0 0.0048027147543415524 0.9952088 0.00479181 1.225549E-10 0 1 2 +22 0 0 0.00140569018309729 0.9985953 0.0014052592 1.0557051E-11 0 1 2 +23 0 0 0.025959029382547023 0.974375 0.025625352 4.9580686E-09 0 1 2 +24 0 0 0.035062818553545601 0.96554476 0.0344552 6.860606E-09 0 1 2 +25 0 0 0.0435714563066327 0.95736414 0.042635527 3.4239802E-09 0 1 2 +26 0 0 0.014657948261883724 0.98544896 0.014551509 1.3158941E-09 0 1 2 +27 0 0 0.0083618103860956319 0.99167305 0.008327509 9.574487E-11 0 1 2 +28 0 0 0.0081901648464538963 0.9918433 0.008155996 7.816004E-11 0 1 2 +29 0 0 0.025003426094698581 0.9753066 0.024693534 3.0869016E-09 0 1 2 +30 0 0 0.032958997100846141 0.96757823 0.032421295 3.7381946E-09 0 1 2 +31 0 0 0.011100762367249332 0.9889606 0.011037588 2.2150382E-10 0 1 2 +32 0 0 0.001602144297411704 0.99839914 0.0016007263 3.5017909E-12 0 1 2 +33 0 0 0.00091552908011568914 0.9990849 0.0009148209 7.622556E-13 0 1 2 +34 0 0 0.024698270577283184 0.97560424 0.024394816 9.156438E-10 0 1 2 +35 0 0 0.0077877307214346608 0.9922425 0.007757637 9.0569406E-11 0 1 2 +36 0 0 0.0046844358446896913 0.9953265 0.004674313 1.07777615E-11 0 1 2 +37 0 0 0.024698270577283184 0.97560424 0.024394816 9.156438E-10 0 1 2 +38 0 0 0.01765443292184056 0.9825005 0.017499987 2.408446E-09 0 1 2 +39 0 0 0.01095962001675865 0.9891002 0.01089918 2.0268184E-10 0 1 2 +40 0 0 0.0045871881228086026 0.9954233 0.004576382 6.802108E-11 0 1 2 +41 0 0 0.11641690690510131 0.89010406 0.109895095 7.33158E-08 0 1 2 +42 0 0 0.010204286632554432 0.9898476 0.010152256 9.365891E-10 0 1 2 +43 0 0 0.011142711158630041 0.98891914 0.01108102 2.055959E-09 0 1 2 +44 0 0 0.011913290735280963 0.9881574 0.011842146 1.1434037E-09 0 1 2 +45 0 0 0.024089826265429194 0.976198 0.023801321 2.4710503E-09 0 1 2 +46 0 0 0.0049056737103747861 0.99510634 0.0048929513 5.991218E-11 0 1 2 +47 0 0 0.013824452798308497 0.98627067 0.01372948 1.058424E-09 0 1 2 +48 0 0 0.0048446397513665267 0.9951671 0.0048325625 2.8125514E-11 0 1 2 +49 0 0 0.01068685349281682 0.98937005 0.0106304465 2.1819495E-10 0 1 2 +50 1 1 0.010307321484771452 0.9897456 0.0066921716 0.0035611067 1 2 0 +51 1 1 0.033012407363040078 0.96752656 0.02600656 0.0064676455 1 2 0 +52 1 1 0.037960064013627892 0.9627514 0.03579031 0.0014575779 1 2 0 +53 1 1 0.11583328786550111 0.8906237 0.10709783 0.0022779782 1 2 0 +54 1 1 0.065635657500629663 0.936472 0.061990798 0.0015358254 1 2 0 +55 1 1 0.15934372439907335 0.8527032 0.1453504 0.0019471206 1 2 0 +56 1 1 0.098892216640845645 0.90584034 0.08973586 0.0044248756 1 2 0 +57 1 1 0.036201489056084446 0.96444595 0.026659163 0.0088957185 1 0 2 +58 1 1 0.01831433568022731 0.98185235 0.016035324 0.002113003 1 2 0 +59 1 1 0.13633346332433335 0.8725516 0.11827276 0.009174371 1 2 0 +60 1 1 0.037609091751875177 0.96308935 0.03205739 0.004853235 1 2 0 +61 1 1 0.059661896677204285 0.942083 0.048887465 0.009028244 1 2 0 +62 1 1 0.010932924536648346 0.9891266 0.008992977 0.0018793689 1 2 0 +63 1 1 0.1448860271834107 0.8651209 0.13346551 0.0014131966 1 2 0 +64 1 1 0.048488610138855845 0.9526682 0.041682206 0.0056510344 1 0 2 +65 1 1 0.012752559180018301 0.9873284 0.006654496 0.006017411 1 0 2 +66 1 1 0.32437517267349469 0.72297895 0.27414086 0.0028792499 1 2 0 +67 1 1 0.014152544366119003 0.98594713 0.008474232 0.005579778 1 2 0 +68 1 1 0.29913708418651769 0.74145776 0.25822604 0.0003164553 1 2 0 +69 1 1 0.022277803774763228 0.9779685 0.016210856 0.0058211614 1 2 0 +70 1 2 0.86252211641638088 0.57672143 0.42209616 0.001182573 2 1 0 +71 1 1 0.017951984596790243 0.9822082 0.009733042 0.008057902 1 0 2 +72 1 1 0.5103301934951745 0.60029733 0.39952287 0.00017800763 1 2 0 +73 1 1 0.073650368218172196 0.92899644 0.06985277 0.0011515534 1 2 0 +74 1 1 0.014289845278877916 0.98581177 0.008972383 0.005214931 1 2 0 +75 1 1 0.014881161940909169 0.985229 0.009711901 0.0050603687 1 2 0 +76 1 1 0.038799306519908344 0.96194375 0.03719588 0.0008600391 1 2 0 +77 1 1 0.25716021132199368 0.7732443 0.2260887 0.0006657438 1 2 0 +78 1 1 0.14286805957946538 0.86686844 0.13055693 0.0025754895 1 2 0 +79 1 1 0.026664715915927045 0.97368765 0.02494495 0.0013670862 1 0 2 +80 1 1 0.024108876525694791 0.9761794 0.017860023 0.005959036 1 2 0 +81 1 1 0.01597211702271948 0.98415476 0.008084919 0.007762048 1 0 2 +82 1 1 0.020076735361690724 0.98012346 0.010045679 0.009831256 1 0 2 +83 1 2 1.5089301971332096 0.7787889 0.22114643 6.3689586E-05 2 1 0 +84 1 1 0.51219176115903042 0.5991809 0.39841336 0.002406612 1 2 0 +85 1 1 0.094871550068080526 0.90948975 0.07977255 0.010739437 1 2 0 +86 1 1 0.032588961436346382 0.96793634 0.029385215 0.0026766758 1 2 0 +87 1 1 0.057693512172543704 0.9439392 0.055348806 0.0007114502 1 2 0 +88 1 1 0.044856999113026041 0.9561342 0.031375647 0.012489466 1 2 0 +89 1 1 0.081623051679496975 0.9216193 0.07428871 0.0040936833 1 2 0 +90 1 1 0.17732237198367184 0.83750975 0.16099443 0.0014946358 1 2 0 +91 1 1 0.085346196629971588 0.91819435 0.0791445 0.0026600028 1 2 0 +92 1 1 0.023456550648371967 0.9768164 0.017536609 0.0056453245 1 2 0 +93 1 1 0.028970723245542097 0.9714449 0.020290121 0.008264454 1 0 2 +94 1 1 0.087882981563431148 0.91586804 0.08024623 0.003884399 1 2 0 +95 1 1 0.032085432800653201 0.96842384 0.022214318 0.00935996 1 2 0 +96 1 1 0.050395938246597986 0.9508529 0.042172417 0.0069737583 1 2 0 +97 1 1 0.021072381409417223 0.9791481 0.015627518 0.005223911 1 2 0 +98 1 1 0.086358994837316017 0.9172649 0.08074706 0.0019872745 1 0 2 +99 1 1 0.043498989321880621 0.9574335 0.035394125 0.0071709203 1 2 0 +100 2 2 0.0011253129694704825 0.9988753 0.0011242428 1.1564506E-07 2 1 0 +101 2 2 0.040019973135030099 0.96077025 0.03921765 1.1391113E-05 2 1 0 +102 2 2 0.052360548423487648 0.94898665 0.051010333 2.9822752E-06 2 1 0 +103 2 2 0.056476572865404355 0.9450886 0.05490551 6.142361E-06 2 1 0 +104 2 2 0.0091857704417590124 0.9908563 0.009142892 7.379481E-07 2 1 0 +105 2 2 0.014972700003367621 0.98513883 0.014861965 1.0640279E-07 2 1 0 +106 2 2 0.052852650255712383 0.94851977 0.05142669 5.3088323E-05 2 1 0 +107 2 2 0.064285257313123734 0.93773746 0.06226263 8.3501254E-07 2 1 0 +108 2 2 0.036701220902616828 0.9639641 0.03603608 7.256182E-07 2 1 0 +109 2 2 0.017375952834142096 0.98277414 0.017222695 2.899598E-06 2 1 0 +110 2 2 0.40617370069911779 0.66619444 0.33343077 0.00037362365 2 1 0 +111 2 2 0.097599022765379448 0.9070125 0.0929744 1.4535398E-05 2 1 0 +112 2 2 0.10143276983895463 0.9035419 0.096438296 1.8730723E-05 2 1 0 +113 2 2 0.018951408588991213 0.98122704 0.018767871 4.2423317E-06 2 1 0 +114 2 2 0.0050048693274474714 0.99500763 0.0049899984 1.9100253E-06 2 1 0 +115 2 2 0.043437981670942004 0.95749193 0.04248243 2.6396723E-05 2 1 0 +116 2 2 0.16752241046970612 0.84575766 0.15421322 3.0336812E-05 2 1 0 +117 2 2 0.041770925874962929 0.95908946 0.040908758 1.9711408E-06 2 1 0 +118 2 2 0.0011367700123326584 0.9988639 0.0011371002 1.1028184E-09 2 1 0 +119 2 2 0.21701051211212155 0.8049215 0.19506057 1.902177E-05 2 1 0 +120 2 2 0.038440730629811432 0.96228874 0.03770575 6.9925945E-06 2 1 0 +121 2 2 0.037864664123807103 0.96284324 0.037132367 2.5961173E-05 2 1 0 +122 2 2 0.01436234244500512 0.9857403 0.014260588 4.3499693E-08 2 1 0 +123 2 2 0.44808955374992387 0.6388475 0.36096635 0.00018588797 2 1 0 +124 2 2 0.066169743843233181 0.935972 0.06401352 1.5777965E-05 2 1 0 +125 2 2 0.2526970107586809 0.7767032 0.22328086 1.6773914E-05 2 1 0 +126 2 2 0.56779862124056424 0.56677175 0.43283176 0.00039698367 2 1 0 +127 2 2 0.46234508080374481 0.62980497 0.36975628 0.0004395013 2 1 0 +128 2 2 0.01575471481712629 0.98436874 0.015629306 1.3211622E-06 2 1 0 +129 2 1 0.71149441960784132 0.50905 0.49091002 3.9881248E-05 1 2 0 +130 2 2 0.093170167083788336 0.91103846 0.08896046 1.632669E-06 2 1 0 +131 2 2 0.46365736580853845 0.628979 0.37097695 4.3197317E-05 2 1 0 +132 2 2 0.0099792246221438764 0.9900704 0.009930234 8.394135E-07 2 1 0 +133 2 1 0.82217403963248004 0.56031615 0.43947518 0.00020995998 1 2 0 +134 2 2 0.10812757202505351 0.8975131 0.102480784 5.052308E-06 2 1 0 +135 2 2 0.052467328962697665 0.9488853 0.051112007 1.6083872E-06 2 1 0 +136 2 2 0.0099868703600679443 0.99006283 0.009934592 4.4171784E-06 2 1 0 +137 2 2 0.15548322444185292 0.85600144 0.14396037 3.7479545E-05 2 1 0 +138 2 2 0.50120926215640438 0.60579765 0.39357048 0.00063224003 2 1 0 +139 2 2 0.2255859111403673 0.7980485 0.20188296 6.9243266E-05 2 1 0 +140 2 2 0.016796130917139922 0.98334414 0.016652549 3.1806762E-06 2 1 0 +141 2 2 0.27604677305242886 0.75877744 0.2410193 0.00020148105 2 1 0 +142 2 2 0.040019973135030099 0.96077025 0.03921765 1.1391113E-05 2 1 0 +143 2 2 0.013687155339038522 0.9864061 0.0135924 1.3977802E-06 2 1 0 +144 2 2 0.010833741704283761 0.98922473 0.010773248 2.6553757E-06 2 1 0 +145 2 2 0.0964507807946485 0.9080546 0.09189998 4.368884E-05 2 1 0 +146 2 2 0.15127181068736736 0.859614 0.14035329 3.09263E-05 2 1 0 +147 2 2 0.20624921409622274 0.8136303 0.18628152 8.931429E-05 2 1 0 +148 2 2 0.025451673692827314 0.9748695 0.025110643 2.0306428E-05 2 1 0 +149 2 2 0.14498138583536518 0.8650384 0.13487196 8.92741E-05 2 1 0 diff --git a/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/MulticlassLogisticRegression-CV-iris.txt b/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/MulticlassLogisticRegression-CV-iris.txt index a077f1e9b6..eabf5f533c 100644 --- a/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/MulticlassLogisticRegression-CV-iris.txt +++ b/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/MulticlassLogisticRegression-CV-iris.txt @@ -1,151 +1,151 @@ Instance Label Assigned Log-loss #1 Score #2 Score #3 Score #1 Class #2 Class #3 Class -5 0 0 0.0039186265649996796 0.99608904 0.003910809 2.0865299E-11 0 1 2 -6 0 0 0.0066394791182328066 0.9933825 0.006617098 2.152395E-10 0 1 2 -8 0 0 0.025494595703479419 0.97482765 0.025172155 2.838915E-09 0 1 2 -9 0 0 0.019190772567240676 0.9809922 0.01900817 2.7058017E-10 0 1 2 -10 0 0 0.0035562493991592786 0.99645007 0.003550249 4.9272335E-12 0 1 2 -11 0 0 0.012517929979546309 0.9875601 0.0124394335 2.8008076E-10 0 1 2 -18 0 0 0.0047769616990695207 0.99523443 0.004765985 8.127231E-12 0 1 2 -20 0 0 0.014707546996037419 0.9854001 0.014600216 8.914893E-11 0 1 2 -21 0 0 0.003647833406238951 0.9963588 0.003641376 3.5084515E-11 0 1 2 -25 0 0 0.032317619176197167 0.968199 0.03180108 1.10023E-09 0 1 2 -28 0 0 0.0057745087199158104 0.99424213 0.0057580057 2.0235997E-11 0 1 2 -31 0 0 0.0071621105155360203 0.9928635 0.0071367947 6.242954E-11 0 1 2 -32 0 0 0.0014066451993830036 0.99859434 0.0014055507 7.3242096E-13 0 1 2 -35 0 0 0.0052270768592320212 0.99478656 0.0052133226 2.5026661E-11 0 1 2 -37 0 0 0.019190772567240676 0.9809922 0.01900817 2.7058017E-10 0 1 2 -40 0 0 0.0033063652339419308 0.9966991 0.003300911 1.895811E-11 0 1 2 -41 0 0 0.077088846429646946 0.9258076 0.0741926 3.3343277E-08 0 1 2 -44 0 0 0.010642032182541446 0.9894144 0.0105858175 3.5036918E-10 0 1 2 -45 0 0 0.017220520546649732 0.9829269 0.017073102 8.5334684E-10 0 1 2 -46 0 0 0.0041847046934752278 0.99582404 0.0041763177 1.5150974E-11 0 1 2 -48 0 0 0.0036583023932260659 0.9963484 0.0036516767 6.700264E-12 0 1 2 -50 1 1 0.0098977138531054429 0.9901511 0.005578102 0.0042707166 1 2 0 -51 1 1 0.031642446182477788 0.96885294 0.024089899 0.007057201 1 2 0 -52 1 1 0.03573585794885261 0.9648951 0.033440154 0.001664746 1 2 0 -54 1 1 0.072409637391112461 0.9301498 0.06803778 0.0018123286 1 2 0 -56 1 1 0.093649650619783725 0.91060174 0.08499099 0.0044072797 1 2 0 -60 1 1 0.042489542543681612 0.9584005 0.03626706 0.005332458 1 2 0 -63 1 1 0.137502875229945 0.87153184 0.1271067 0.0013614136 1 2 0 -64 1 1 0.054985501307998973 0.9464989 0.048023954 0.0054772357 1 0 2 -66 1 1 0.32485007538570837 0.7226357 0.27483696 0.0025274218 1 2 0 -68 1 1 0.41495870735351531 0.66036755 0.33927307 0.00035943443 1 2 0 -69 1 1 0.021870392416744027 0.97836703 0.015370483 0.006262353 1 2 0 -70 1 2 0.95191086445840967 0.61301166 0.38600272 0.0009856796 2 1 0 -71 1 1 0.019884826671693976 0.9803116 0.011781351 0.007907056 1 0 2 -72 1 1 0.59602090426171883 0.55099976 0.44882512 0.00017505078 1 2 0 -73 1 1 0.061366106483072287 0.94047886 0.058446974 0.001074166 1 2 0 -74 1 1 0.01440358171033043 0.98569965 0.008161617 0.0061388216 1 2 0 -76 1 1 0.038602842055599461 0.96213275 0.036847353 0.0010197883 1 2 0 -77 1 1 0.29142764060847787 0.7471961 0.25208423 0.00071971776 1 2 0 -79 1 1 0.031789553326302268 0.9687104 0.030109784 0.0011796883 1 0 2 -82 1 1 0.020876629838716645 0.9793398 0.011350486 0.009309866 1 0 2 -88 1 1 0.040395748253186578 0.9604093 0.027647337 0.01194328 1 2 0 -90 1 1 0.166073088823076 0.8469843 0.15168492 0.0013306275 1 2 0 -91 1 1 0.078327406928112728 0.92466164 0.07274388 0.0025945483 1 2 0 -92 1 1 0.023790137724984652 0.9764906 0.017208781 0.006300524 1 2 0 -93 1 1 0.031245715484219467 0.9692374 0.022426113 0.0083365 1 0 2 -95 1 1 0.026741054389737642 0.9736133 0.017711794 0.008674911 1 2 0 -96 1 1 0.04639434288106066 0.9546654 0.038535807 0.006798673 1 2 0 -97 1 1 0.020153545641826632 0.9800482 0.014183712 0.005768105 1 2 0 -98 1 1 0.10663172345345885 0.89885664 0.09910672 0.0020366241 1 0 2 -99 1 1 0.042470449845490664 0.9584188 0.034200437 0.0073807975 1 2 0 -100 2 2 0.00061608453557407644 0.9993841 0.0006157393 5.2001777E-08 2 1 0 -102 2 2 0.035200613028489182 0.9654117 0.03458606 2.1171227E-06 2 1 0 -104 2 2 0.0055899402333488843 0.99442565 0.005573957 4.1343924E-07 2 1 0 -105 2 2 0.010776984144366402 0.9892809 0.010718968 7.398926E-08 2 1 0 -106 2 2 0.035451432887178884 0.9651696 0.03480039 2.9798488E-05 2 1 0 -108 2 2 0.02539346900894791 0.97492623 0.025073528 4.914205E-07 2 1 0 -109 2 2 0.010955823550611778 0.989104 0.010894373 1.8228092E-06 2 1 0 -111 2 2 0.064874652452474105 0.93718493 0.06280472 1.0175524E-05 2 1 0 -112 2 2 0.065967573668299667 0.9361612 0.06382524 1.3542564E-05 2 1 0 -113 2 2 0.010099395894045529 0.98995143 0.010046235 2.2420422E-06 2 1 0 -115 2 2 0.025224193337550441 0.9750913 0.024892632 1.615105E-05 2 1 0 -117 2 2 0.040344238358527816 0.96045876 0.039539583 1.5799143E-06 2 1 0 -120 2 2 0.02318047647021565 0.9770861 0.022909554 4.537731E-06 2 1 0 -121 2 2 0.023067445323975651 0.9771966 0.022788834 1.4752109E-05 2 1 0 -122 2 2 0.010407897593236567 0.9896461 0.010353751 3.0826786E-08 2 1 0 -123 2 2 0.3268478109707687 0.7211935 0.27864572 0.00016074626 2 1 0 -125 2 2 0.24618388448271417 0.78177845 0.21820657 1.498082E-05 2 1 0 -128 2 2 0.0093602944019322582 0.9906834 0.009315706 7.625066E-07 2 1 0 -129 2 1 0.72903371465671218 0.517586 0.48237488 3.915742E-05 1 2 0 -131 2 2 0.50473360924795707 0.60366637 0.3962909 4.2747015E-05 2 1 0 -132 2 2 0.0054111588821948668 0.99460346 0.0053958558 4.5496841E-07 2 1 0 -133 2 1 0.80138632910758123 0.5511005 0.44870648 0.00019297059 1 2 0 -137 2 2 0.13871576668047261 0.8704754 0.12949541 2.8931609E-05 2 1 0 -138 2 2 0.41129950849971153 0.6627884 0.33667925 0.0005324046 2 1 0 -141 2 2 0.15623071913140255 0.8553618 0.14447537 0.00016272673 2 1 0 -144 2 2 0.0057010728771068496 0.99431515 0.005683601 1.4550411E-06 2 1 0 -145 2 2 0.050842734596971917 0.9504281 0.04954274 2.9357496E-05 2 1 0 -147 2 2 0.1444256224627285 0.8655193 0.13441183 6.885606E-05 2 1 0 -0 0 0 0.0038471219092532103 0.99616027 0.0038398658 4.272304E-13 0 1 2 -1 0 0 0.012569414513148094 0.98750925 0.012490312 1.16851225E-11 0 1 2 -2 0 0 0.0067682510684477139 0.9932546 0.0067451415 3.889237E-12 0 1 2 -3 0 0 0.017426475046109894 0.9827245 0.017275961 4.7416498E-11 0 1 2 -4 0 0 0.0033359676846921181 0.9966696 0.003330729 3.6924307E-13 0 1 2 -7 0 0 0.0071029796876616886 0.9929222 0.0070774616 2.3269897E-12 0 1 2 -12 0 0 0.013135919395226348 0.98695 0.013049681 9.69176E-12 0 1 2 -13 0 0 0.006740587102816476 0.9932821 0.006718063 5.392806E-12 0 1 2 -14 0 0 0.00042721812913112326 0.9995729 0.00042694033 4.819717E-16 0 1 2 -15 0 0 0.00056550992629539202 0.99943465 0.00056577666 3.0879463E-15 0 1 2 -16 0 0 0.0010137330102610047 0.9989868 0.0010132674 2.4230277E-14 0 1 2 -17 0 0 0.003937715283349302 0.99607 0.0039304383 7.4476986E-13 0 1 2 -19 0 0 0.00293255272107629 0.99707174 0.0029287196 3.5780553E-13 0 1 2 -22 0 0 0.0011405890557899762 0.99886006 0.0011397742 6.27105E-14 0 1 2 -23 0 0 0.01713659836841247 0.9830094 0.01699105 7.186254E-11 0 1 2 -24 0 0 0.030869489575278578 0.9696021 0.030398024 1.229949E-10 0 1 2 -26 0 0 0.010402657749792236 0.98965126 0.010348479 1.5896721E-11 0 1 2 -27 0 0 0.0050296697692560326 0.99498296 0.0050173127 6.650896E-13 0 1 2 -29 0 0 0.018460284327568528 0.98170906 0.018290574 4.4078418E-11 0 1 2 -30 0 0 0.021263117789093924 0.97896135 0.021038212 5.088387E-11 0 1 2 -33 0 0 0.00067328225244051433 0.99932694 0.0006728331 2.645334E-15 0 1 2 -34 0 0 0.013916015058125409 0.98618037 0.013819583 9.011755E-12 0 1 2 -36 0 0 0.0020971218736402456 0.9979051 0.0020952104 4.3312952E-14 0 1 2 -38 0 0 0.012599956415221663 0.9874791 0.012520914 3.2850903E-11 0 1 2 -39 0 0 0.0066385190914096566 0.99338347 0.006616723 1.6090942E-12 0 1 2 -42 0 0 0.0082874029946546218 0.99174684 0.008253521 1.1761102E-11 0 1 2 -43 0 0 0.0088386780169330721 0.99120027 0.0087993285 2.8893552E-11 0 1 2 -47 0 0 0.010122275828759237 0.9899288 0.010071042 1.2646444E-11 0 1 2 -49 0 0 0.0062646587177839679 0.9937549 0.0062455637 1.729813E-12 0 1 2 -53 1 1 0.10698867632136459 0.89853585 0.09846105 0.0030030068 1 2 0 -55 1 1 0.13397672143525333 0.8746104 0.12360578 0.0017838533 1 2 0 -57 1 1 0.034013499575166965 0.96655846 0.030011753 0.0034298701 1 0 2 -58 1 1 0.014321588595529325 0.9857805 0.010956884 0.0032627187 1 2 0 -59 1 1 0.089034322664890295 0.9148142 0.07726592 0.007920011 1 2 0 -61 1 1 0.037737148309352063 0.962966 0.028062146 0.008971874 1 2 0 -62 1 1 0.010682094145451054 0.98937476 0.00659236 0.004032912 1 2 0 -65 1 1 0.013174752701392319 0.98691165 0.010195948 0.002892394 1 0 2 -67 1 1 0.011484394147178454 0.9885813 0.007202252 0.0042163264 1 0 2 -75 1 1 0.013111581404427745 0.986974 0.0077194124 0.005306578 1 0 2 -78 1 1 0.11794544915001207 0.88874453 0.108638965 0.0026165938 1 2 0 -80 1 1 0.019020538797954168 0.9811592 0.010528394 0.008312513 1 2 0 -81 1 1 0.015925180791274182 0.98420095 0.011949363 0.0038497108 1 0 2 -83 1 2 2.0406980454630639 0.87002933 0.12993798 3.258634E-05 2 1 0 -84 1 1 0.43994994621280786 0.64406866 0.35437283 0.0015585137 1 2 0 -85 1 1 0.05211593834564536 0.9492188 0.04295896 0.007822319 1 2 0 -86 1 1 0.024243581990507159 0.97604793 0.020361558 0.0035904064 1 2 0 -87 1 1 0.065390833308409307 0.9367013 0.061896957 0.0014018206 1 2 0 -89 1 1 0.062605639482612899 0.9393138 0.055903245 0.0047830124 1 2 0 -94 1 1 0.064166784413406702 0.93784857 0.058178328 0.003973016 1 2 0 -101 2 2 0.016407972136600949 0.9837259 0.016270882 3.321601E-06 2 1 0 -103 2 2 0.021377102120025193 0.97884977 0.021148358 1.7575348E-06 2 1 0 -107 2 2 0.015548256759951539 0.984572 0.015427698 2.396388E-07 2 1 0 -110 2 2 0.31425472469679994 0.730333 0.26941967 0.0002474155 2 1 0 -114 2 2 0.0015654293382317786 0.9984358 0.0015639706 3.5057428E-07 2 1 0 -116 2 2 0.082808513856507754 0.9205274 0.07945967 1.3067251E-05 2 1 0 -118 2 2 8.392686160427524E-05 0.9999161 8.372341E-05 1.07175394E-10 2 1 0 -119 2 2 0.085332889130497608 0.9182066 0.081782915 1.0257855E-05 2 1 0 -124 2 2 0.03029776689626696 0.9701566 0.02983821 5.02674E-06 2 1 0 -126 2 2 0.42805961427942696 0.65177256 0.34790462 0.00032288703 2 1 0 -127 2 2 0.37273980651049665 0.68884444 0.31086212 0.0002934189 2 1 0 -130 2 2 0.023508540435441566 0.97676563 0.023233732 5.989922E-07 2 1 0 -134 2 2 0.039031011947558578 0.9617209 0.038277302 1.6219885E-06 2 1 0 -135 2 2 0.01278401224870694 0.98729736 0.012701993 5.567945E-07 2 1 0 -136 2 2 0.0040655412946085491 0.9959427 0.0040563564 8.3937186E-07 2 1 0 -139 2 2 0.12014629655188541 0.8867907 0.11317021 3.931525E-05 2 1 0 -140 2 2 0.0054330927887431615 0.99458164 0.0054175323 7.817332E-07 2 1 0 -142 2 2 0.016407972136600949 0.9837259 0.016270882 3.321601E-06 2 1 0 -143 2 2 0.0040641049565237689 0.99594414 0.00405544 2.884136E-07 2 1 0 -146 2 2 0.063336850854471993 0.93862724 0.061356686 1.6151482E-05 2 1 0 -148 2 2 0.013229351377581255 0.9868578 0.013137038 5.0962526E-06 2 1 0 -149 2 2 0.090927626330238626 0.9130838 0.086879514 3.6539095E-05 2 1 0 +5 0 0 0.0039186265649996796 0.99608904 0.003910874 2.0866812E-11 0 1 2 +6 0 0 0.0066394791182328066 0.9933825 0.006617177 2.1525429E-10 0 1 2 +8 0 0 0.025494595703479419 0.97482765 0.025172275 2.8391045E-09 0 1 2 +9 0 0 0.019190772567240676 0.9809922 0.019008234 2.706003E-10 0 1 2 +10 0 0 0.0035562493991592786 0.99645007 0.0035502762 4.9275905E-12 0 1 2 +11 0 0 0.012517929979546309 0.9875601 0.012439534 2.801027E-10 0 1 2 +18 0 0 0.00477791994040919 0.9952335 0.004766044 8.12785E-12 0 1 2 +20 0 0 0.014707546996037419 0.9854001 0.014600349 8.915574E-11 0 1 2 +21 0 0 0.003647833406238951 0.9963588 0.0036414368 3.5086923E-11 0 1 2 +25 0 0 0.032318111675433796 0.96819854 0.03180129 1.1003118E-09 0 1 2 +28 0 0 0.0057745087199158104 0.99424213 0.0057580387 2.0237386E-11 0 1 2 +31 0 0 0.0071621105155360203 0.9928635 0.0071368967 6.2433704E-11 0 1 2 +32 0 0 0.0014066451993830036 0.99859434 0.0014055514 7.3247544E-13 0 1 2 +35 0 0 0.0052270768592320212 0.99478656 0.005213345 2.5028236E-11 0 1 2 +37 0 0 0.019190772567240676 0.9809922 0.019008234 2.706003E-10 0 1 2 +40 0 0 0.0033063652339419308 0.9966991 0.003300944 1.8959338E-11 0 1 2 +41 0 0 0.077089812148774681 0.9258067 0.07419313 3.3345312E-08 0 1 2 +44 0 0 0.010642032182541446 0.9894144 0.0105860345 3.5039727E-10 0 1 2 +45 0 0 0.017220520546649732 0.9829269 0.017073257 8.534022E-10 0 1 2 +46 0 0 0.0041847046934752278 0.99582404 0.0041763554 1.5152131E-11 0 1 2 +48 0 0 0.0036583023932260659 0.9963484 0.0036516993 6.7007368E-12 0 1 2 +50 1 1 0.0098974730630395254 0.99015135 0.0055783894 0.0042703073 1 2 0 +51 1 1 0.031642938349304373 0.96885246 0.024090974 0.007056518 1 2 0 +52 1 1 0.035737278733103894 0.96489376 0.03344172 0.0016645745 1 2 0 +54 1 1 0.072412264703038826 0.93014735 0.06804041 0.0018121488 1 2 0 +56 1 1 0.093652465245797983 0.9105992 0.08499416 0.0044068047 1 2 0 +60 1 1 0.042491221723503017 0.9583989 0.03626898 0.0053321123 1 2 0 +63 1 1 0.13750937236448643 0.8715262 0.12711242 0.001361274 1 2 0 +64 1 1 0.054981974780578077 0.9465022 0.04802037 0.005477439 1 0 2 +66 1 1 0.32486129303978856 0.7226276 0.27484533 0.0025271494 1 2 0 +68 1 1 0.41497323928737251 0.66035795 0.33928275 0.00035939485 1 2 0 +69 1 1 0.021870636107098145 0.9783668 0.015371303 0.006261893 1 2 0 +70 1 2 0.95192823630695889 0.6130184 0.385996 0.000985549 2 1 0 +71 1 1 0.019884097051136185 0.9803123 0.011780373 0.007907392 1 0 2 +72 1 1 0.59604091691945504 0.55098873 0.44883615 0.00017502924 1 2 0 +73 1 1 0.061369909104942041 0.9404753 0.058450583 0.0010740681 1 2 0 +74 1 1 0.01440358171033043 0.98569965 0.008162019 0.0061382917 1 2 0 +76 1 1 0.038604514721613192 0.96213114 0.036849145 0.0010196886 1 2 0 +77 1 1 0.29143689409760176 0.74718916 0.25209117 0.0007196292 1 2 0 +79 1 1 0.031787645901562245 0.9687123 0.03010791 0.0011797558 1 0 2 +82 1 1 0.020876142942286392 0.97934026 0.011349594 0.009310292 1 0 2 +88 1 1 0.040395996500067109 0.96040905 0.02764859 0.011942271 1 2 0 +90 1 1 0.16608146321939379 0.84697723 0.15169221 0.0013305076 1 2 0 +91 1 1 0.078330758907475886 0.92465854 0.07274721 0.0025942985 1 2 0 +92 1 1 0.023790381883611583 0.9764904 0.01720961 0.0063000256 1 2 0 +93 1 1 0.031244731541730415 0.96923834 0.022424724 0.0083369175 1 0 2 +95 1 1 0.026741238049869182 0.97361314 0.017712766 0.008674208 1 2 0 +96 1 1 0.046395529148936118 0.9546643 0.038537607 0.006798083 1 2 0 +97 1 1 0.020153788914160642 0.98004794 0.014184415 0.005767607 1 2 0 +98 1 1 0.106625556491314 0.8988622 0.09910103 0.0020367077 1 0 2 +99 1 1 0.042471382705010136 0.9584179 0.03420201 0.007380164 1 2 0 +100 2 2 0.00061608453557407644 0.9993841 0.00061573554 5.199305E-08 2 1 0 +102 2 2 0.035199625186905135 0.9654127 0.03458504 2.1167655E-06 2 1 0 +104 2 2 0.0055899402333488843 0.99442565 0.0055738376 4.1337103E-07 2 1 0 +105 2 2 0.010776502140667386 0.98928136 0.010718488 7.397515E-08 2 1 0 +106 2 2 0.035450012508994544 0.965171 0.034799412 2.9794508E-05 2 1 0 +108 2 2 0.025392551945450037 0.9749271 0.025072392 4.9133615E-07 2 1 0 +109 2 2 0.010955823550611778 0.989104 0.010894274 1.8224963E-06 2 1 0 +111 2 2 0.06487274446443414 0.9371867 0.062802896 1.01739515E-05 2 1 0 +112 2 2 0.065966109277555141 0.9361626 0.06382388 1.3540421E-05 2 1 0 +113 2 2 0.010098914216835852 0.9899519 0.010046053 2.241711E-06 2 1 0 +115 2 2 0.025224193337550441 0.9750913 0.024892442 1.6148571E-05 2 1 0 +117 2 2 0.040342314546391055 0.9604606 0.039537944 1.5796084E-06 2 1 0 +120 2 2 0.02318047647021565 0.9770861 0.02290926 4.5370043E-06 2 1 0 +121 2 2 0.023067445323975651 0.9771966 0.022788497 1.4749999E-05 2 1 0 +122 2 2 0.010407475995621918 0.9896465 0.0103532225 3.0820733E-08 2 1 0 +123 2 2 0.32683946363533112 0.7211995 0.27863985 0.00016072435 2 1 0 +125 2 2 0.24617290564122887 0.78178704 0.21819793 1.4978335E-05 2 1 0 +128 2 2 0.0093598130805993083 0.99068385 0.009315497 7.6238445E-07 2 1 0 +129 2 1 0.72900344169343156 0.51757145 0.48238948 3.9151782E-05 1 2 0 +131 2 2 0.50471494799167183 0.60367763 0.39627966 4.27398E-05 2 1 0 +132 2 2 0.0054106794579163063 0.99460393 0.0053957785 4.5489728E-07 2 1 0 +133 2 1 0.80135750397993166 0.5510876 0.4487194 0.00019294629 1 2 0 +137 2 2 0.13871001490885845 0.8704804 0.12949048 2.8926946E-05 2 1 0 +138 2 2 0.41129015581014772 0.6627946 0.336673 0.00053233345 2 1 0 +141 2 2 0.15623044039725695 0.85536206 0.14447519 0.00016270345 2 1 0 +144 2 2 0.0057010728771068496 0.99431515 0.0056835874 1.4548094E-06 2 1 0 +145 2 2 0.050842734596971917 0.9504281 0.049542528 2.9353128E-05 2 1 0 +147 2 2 0.14442252350825718 0.86552197 0.13440919 6.884588E-05 2 1 0 +0 0 0 0.0039052826303189398 0.99610233 0.0038976613 4.373769E-13 0 1 2 +1 0 0 0.01266768309090782 0.9874122 0.012587332 1.2107598E-11 0 1 2 +2 0 0 0.0068416452974198244 0.9931817 0.006818552 4.000567E-12 0 1 2 +3 0 0 0.017726871077341992 0.9824293 0.017570425 4.9032733E-11 0 1 2 +4 0 0 0.0033979263574399499 0.99660784 0.003391916 3.7693964E-13 0 1 2 +7 0 0 0.0072298301139198445 0.99279624 0.0072038285 2.3913504E-12 0 1 2 +12 0 0 0.013244632293206767 0.9868427 0.013157243 1.0024197E-11 0 1 2 +13 0 0 0.0067710715157424523 0.9932518 0.006747837 5.5424797E-12 0 1 2 +14 0 0 0.00043103446373494287 0.99956906 0.00043064685 4.8664584E-16 0 1 2 +15 0 0 0.00058268592186393035 0.9994175 0.0005829225 3.1098986E-15 0 1 2 +16 0 0 0.0010309167062990282 0.9989696 0.0010306925 2.4610629E-14 0 1 2 +17 0 0 0.0040006687487908078 0.9960073 0.0039928784 7.635827E-13 0 1 2 +19 0 0 0.0030078779737593791 0.99699664 0.0030031486 3.6465668E-13 0 1 2 +22 0 0 0.001151091500502608 0.9988496 0.0011507738 6.352539E-14 0 1 2 +23 0 0 0.017549485773855872 0.9826036 0.017396389 7.461256E-11 0 1 2 +24 0 0 0.031930897469959613 0.9685735 0.031426914 1.2703529E-10 0 1 2 +26 0 0 0.010642996060550395 0.98941344 0.010586246 1.640988E-11 0 1 2 +27 0 0 0.0051164761595189259 0.9948966 0.005103707 6.8218895E-13 0 1 2 +29 0 0 0.018854160559333675 0.98132247 0.018677382 4.5549585E-11 0 1 2 +30 0 0 0.021645552448278319 0.97858703 0.021413254 5.2732024E-11 0 1 2 +33 0 0 0.00068855143500750813 0.9993117 0.00068878755 2.6637945E-15 0 1 2 +34 0 0 0.014088646318605779 0.98601013 0.01399014 9.314752E-12 0 1 2 +36 0 0 0.0021105014325534205 0.9978917 0.0021083653 4.4323682E-14 0 1 2 +38 0 0 0.012737285781813377 0.9873435 0.012656637 3.3930508E-11 0 1 2 +39 0 0 0.0067491682510599756 0.99327356 0.006726153 1.6541226E-12 0 1 2 +42 0 0 0.0084113983536726177 0.9916239 0.00837582 1.2086278E-11 0 1 2 +43 0 0 0.0090751521004575654 0.9909659 0.009034118 2.9838947E-11 0 1 2 +47 0 0 0.010281546721274558 0.9897711 0.010228623 1.3024835E-11 0 1 2 +49 0 0 0.0063428746789303565 0.9936772 0.006322958 1.7793256E-12 0 1 2 +53 1 1 0.10957930479889894 0.8962111 0.10103965 0.0027492002 1 2 0 +55 1 1 0.13379505020802107 0.87476933 0.12363272 0.0015979259 1 2 0 +57 1 1 0.032105436132285683 0.9684045 0.02807166 0.0035239363 1 0 2 +58 1 1 0.014090036677320443 0.98600876 0.011049821 0.0029414892 1 2 0 +59 1 1 0.088947540076524279 0.91489357 0.0779023 0.007204062 1 2 0 +61 1 1 0.036890756538712828 0.9637814 0.028116023 0.008102445 1 2 0 +62 1 1 0.010654201210582049 0.98940235 0.006855015 0.0037426846 1 2 0 +65 1 1 0.012215594827445041 0.9878587 0.009231208 0.0029101213 1 0 2 +67 1 1 0.010909061911610221 0.9891502 0.0065822112 0.0042676125 1 0 2 +75 1 1 0.012421607302850537 0.9876552 0.006991635 0.005353076 1 0 2 +78 1 1 0.11785578577155728 0.8888242 0.10883135 0.0023444106 1 2 0 +80 1 1 0.018669348552792966 0.98150384 0.010812856 0.007683281 1 2 0 +81 1 1 0.015171535262657397 0.984943 0.011096936 0.00396 1 0 2 +83 1 2 2.0409796221221623 0.87007 0.1299014 2.8672015E-05 2 1 0 +84 1 1 0.43536851973424479 0.6470262 0.35158497 0.0013889617 1 2 0 +85 1 1 0.050369736069722143 0.9508778 0.042176057 0.0069461297 1 2 0 +86 1 1 0.023870469129919129 0.9764122 0.020375451 0.0032124338 1 2 0 +87 1 1 0.067410212010544465 0.93481165 0.06390876 0.0012797081 1 2 0 +89 1 1 0.063293670471973457 0.9386678 0.056962185 0.0043699616 1 2 0 +94 1 1 0.064283159763128858 0.93773943 0.058663167 0.0035974565 1 2 0 +101 2 2 0.016423241110482573 0.9837109 0.016285963 2.9112189E-06 2 1 0 +103 2 2 0.021625026364756963 0.9786071 0.021391531 1.5351787E-06 2 1 0 +107 2 2 0.015677938925009325 0.9844443 0.015555596 2.0633152E-07 2 1 0 +110 2 2 0.31733201991188631 0.728089 0.2716928 0.00021816233 2 1 0 +114 2 2 0.0015687724332034386 0.99843246 0.001567381 3.0577573E-07 2 1 0 +116 2 2 0.083694952996644006 0.91971177 0.08027698 1.1447655E-05 2 1 0 +118 2 2 8.2973107700441223E-05 0.99991703 8.340552E-05 8.99174E-11 2 1 0 +119 2 2 0.083391697723947486 0.9199907 0.08000028 8.9567075E-06 2 1 0 +124 2 2 0.030993488916279836 0.9694819 0.03051392 4.4014096E-06 2 1 0 +126 2 2 0.42598822225339628 0.65312403 0.34659016 0.00028579155 2 1 0 +127 2 2 0.37475535349032807 0.68745744 0.31228298 0.00025961632 2 1 0 +130 2 2 0.023461371183951175 0.9768117 0.023187635 5.152995E-07 2 1 0 +134 2 2 0.039185036833512002 0.96157277 0.038425777 1.4153605E-06 2 1 0 +135 2 2 0.012777794001199237 0.9873035 0.012695913 4.7814564E-07 2 1 0 +136 2 2 0.0041899719117273965 0.9958188 0.004180388 7.360644E-07 2 1 0 +139 2 2 0.12075187531073932 0.88625383 0.113711946 3.4401823E-05 2 1 0 +140 2 2 0.0054893680446388747 0.9945257 0.0054734107 6.795554E-07 2 1 0 +142 2 2 0.016423241110482573 0.9837109 0.016285963 2.9112189E-06 2 1 0 +143 2 2 0.0041432263192192472 0.99586535 0.004134322 2.5040606E-07 2 1 0 +146 2 2 0.062243437538287393 0.9396541 0.060331948 1.4102009E-05 2 1 0 +148 2 2 0.013613196562563339 0.98647904 0.013516276 4.4905573E-06 2 1 0 +149 2 2 0.092099668430092721 0.91201425 0.08795362 3.2270033E-05 2 1 0 diff --git a/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/MulticlassLogisticRegression-TrainTest-iris.txt b/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/MulticlassLogisticRegression-TrainTest-iris.txt index 1f136318b1..74dc392a4f 100644 --- a/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/MulticlassLogisticRegression-TrainTest-iris.txt +++ b/test/BaselineOutput/Common/MulticlassLogisticRegression/netcoreapp/MulticlassLogisticRegression-TrainTest-iris.txt @@ -1,151 +1,151 @@ Instance Label Assigned Log-loss #1 Score #2 Score #3 Score #1 Class #2 Class #3 Class -0 0 0 0.0011148704666220459 0.99888575 0.0011145204 1.3776857E-13 0 1 2 -1 0 0 0.0070600595001449043 0.9929648 0.007035469 5.980848E-12 0 1 2 -2 0 0 0.0023107383603406979 0.9976919 0.002307759 1.30608E-12 0 1 2 -3 0 0 0.0071973505498515739 0.9928285 0.0071717617 1.5530932E-11 0 1 2 -4 0 0 0.00077247645619236299 0.9992278 0.0007718982 9.492881E-14 0 1 2 -5 0 0 0.00070763824108637647 0.9992926 0.00070762733 1.7460614E-13 0 1 2 -6 0 0 0.0015249548965825181 0.9984762 0.0015239433 1.9716312E-12 0 1 2 -7 0 0 0.0023698852189009997 0.9976329 0.0023674963 7.53184E-13 0 1 2 -8 0 0 0.01027010486628236 0.98978245 0.010217615 4.4595366E-11 0 1 2 -9 0 0 0.0076313190029468729 0.9923977 0.0076022604 3.5732832E-12 0 1 2 -10 0 0 0.00077915735755335308 0.99922115 0.0007790165 3.6408105E-14 0 1 2 -11 0 0 0.0034885388576309144 0.99651754 0.0034823413 2.83667E-12 0 1 2 -12 0 0 0.0075292801506795124 0.992499 0.0075008543 4.1048614E-12 0 1 2 -13 0 0 0.0024023278603287428 0.99760056 0.0023991016 1.5549362E-12 0 1 2 -14 0 0 8.1065602621708377E-05 0.99991894 8.0906575E-05 1.6902284E-16 0 1 2 -15 0 0 5.149973917474053E-05 0.9999485 5.1647687E-05 6.3881086E-16 0 1 2 -16 0 0 0.00015259953334155612 0.9998474 0.00015233555 7.692787E-15 0 1 2 -17 0 0 0.0010471460236901183 0.9989534 0.0010464673 2.6370047E-13 0 1 2 -18 0 0 0.0010852740224894391 0.9989153 0.0010850534 7.665227E-14 0 1 2 -19 0 0 0.00050831837325152984 0.9994918 0.00050798786 8.2925454E-14 0 1 2 -20 0 0 0.0050735213310076762 0.9949393 0.005060825 1.1999263E-12 0 1 2 -21 0 0 0.00068950576665922968 0.99931073 0.00068974565 3.0276918E-13 0 1 2 -22 0 0 0.00016690693355841423 0.9998331 0.00016714552 1.24774365E-14 0 1 2 -23 0 0 0.0060968509418567766 0.9939217 0.006078474 3.6398856E-11 0 1 2 -24 0 0 0.011000658927949667 0.9890596 0.010940627 2.9289283E-11 0 1 2 -25 0 0 0.015137646970205936 0.98497635 0.015023758 2.1512653E-11 0 1 2 -26 0 0 0.0030670062937996402 0.9969377 0.003062062 6.018899E-12 0 1 2 -27 0 0 0.0016346217426049894 0.9983667 0.0016332197 2.287227E-13 0 1 2 -28 0 0 0.0016097859543283302 0.9983915 0.0016089765 1.9991067E-13 0 1 2 -29 0 0 0.0072956329818799339 0.9927309 0.007268736 1.3519734E-11 0 1 2 -30 0 0 0.010519988636096064 0.98953515 0.010464923 1.956453E-11 0 1 2 -31 0 0 0.0020780682572110079 0.9979241 0.0020754703 9.251467E-13 0 1 2 -32 0 0 0.0001897991996934539 0.9998102 0.00019020819 2.480733E-15 0 1 2 -33 0 0 8.392686160427524E-05 0.9999161 8.372972E-05 5.0225263E-16 0 1 2 -34 0 0 0.0076313190029468729 0.9923977 0.0076022604 3.5732832E-12 0 1 2 -35 0 0 0.0015668620918511083 0.99843436 0.0015660885 2.637719E-13 0 1 2 -36 0 0 0.00075529720050108832 0.999245 0.00075494894 2.1162744E-14 0 1 2 -37 0 0 0.0076313190029468729 0.9923977 0.0076022604 3.5732832E-12 0 1 2 -38 0 0 0.0048484729816518539 0.99516326 0.0048371274 1.0765147E-11 0 1 2 -39 0 0 0.0023670174117273336 0.9976358 0.0023640736 5.7310146E-13 0 1 2 -40 0 0 0.00071431870928224299 0.99928594 0.0007139834 1.5880794E-13 0 1 2 -41 0 0 0.058616606694306482 0.94306827 0.05693195 1.3674992E-09 0 1 2 -42 0 0 0.0023202972115890941 0.9976824 0.002317787 2.964662E-12 0 1 2 -43 0 0 0.0018701744864162277 0.9981316 0.0018682877 1.1569887E-11 0 1 2 -44 0 0 0.0022153938202876546 0.99778706 0.0022129146 3.5982877E-12 0 1 2 -45 0 0 0.0066394791182328066 0.9933825 0.0066177957 1.5050398E-11 0 1 2 -46 0 0 0.00079442815684458367 0.9992059 0.0007942036 9.45553E-14 0 1 2 -47 0 0 0.0033950556005307988 0.9966107 0.0033896863 3.7430263E-12 0 1 2 -48 0 0 0.00078011177567713298 0.9992202 0.00078014436 4.7848332E-14 0 1 2 -49 0 0 0.0023355318196242821 0.9976672 0.0023323793 6.583122E-13 0 1 2 -50 1 1 0.0060477374096205563 0.9939705 0.0040119863 0.002017312 1 2 0 -51 1 1 0.023358802522272765 0.9769119 0.018574875 0.0045133037 1 2 0 -52 1 1 0.03167006942034075 0.9688262 0.030503416 0.0006706227 1 2 0 -53 1 1 0.085153806889826739 0.918371 0.080720015 0.0009088988 1 2 0 -54 1 1 0.063849951086753426 0.93814576 0.061179116 0.0006753106 1 2 0 -55 1 1 0.089629429284646664 0.9142699 0.08488794 0.00084198435 1 2 0 -56 1 1 0.08483648364691608 0.9186625 0.07830901 0.0030285995 1 2 0 -57 1 1 0.019588462132268222 0.98060215 0.016944302 0.0024533693 1 0 2 -58 1 1 0.0099349166147827024 0.9901143 0.008975584 0.0009100606 1 2 0 -59 1 1 0.092321899949188477 0.9118116 0.08203742 0.006150943 1 2 0 -60 1 1 0.014485823445928967 0.9856186 0.012574332 0.0018072706 1 2 0 -61 1 1 0.045088304308168195 0.95591307 0.037453618 0.006633371 1 2 0 -62 1 1 0.0040779297962239254 0.9959304 0.00350171 0.00056805444 1 2 0 -63 1 1 0.10219954324186892 0.9028494 0.09655303 0.0005976357 1 2 0 -64 1 1 0.04381778370401148 0.95712835 0.040380865 0.0024907354 1 0 2 -65 1 1 0.0080509348386470451 0.9919814 0.004386741 0.0036320223 1 0 2 -66 1 1 0.25712513878794285 0.77327144 0.22504078 0.0016878267 1 2 0 -67 1 1 0.0046958738522607582 0.99531513 0.0024410768 0.002243941 1 0 2 -68 1 1 0.42020607515071323 0.65691143 0.3430131 7.5479635E-05 1 2 0 -69 1 1 0.0088743981357345281 0.99116486 0.006175138 0.002660179 1 2 0 -70 1 2 1.119156119237483 0.67287296 0.32655525 0.00057173334 2 1 0 -71 1 1 0.010574322077426965 0.9894814 0.006254719 0.0042638187 1 0 2 -72 1 1 0.61366271154662444 0.5413644 0.45859504 4.0561474E-05 1 2 0 -73 1 1 0.03377105479336788 0.9667928 0.03281732 0.0003900002 1 2 0 -74 1 1 0.0076222497957970196 0.9924067 0.004714336 0.0028790454 1 2 0 -75 1 1 0.0093469378209280057 0.9906966 0.006279014 0.0030244375 1 2 0 -76 1 1 0.03111313795737956 0.9693659 0.030329114 0.00030520634 1 2 0 -77 1 1 0.36622901478232017 0.693344 0.30640054 0.00025561816 1 2 0 -78 1 1 0.12305449703206547 0.8842155 0.11444254 0.0013420355 1 2 0 -79 1 1 0.017178073477986378 0.9829686 0.016676014 0.00035517136 1 0 2 -80 1 1 0.0099248633048492161 0.9901242 0.007182131 0.002693587 1 2 0 -81 1 1 0.0061249768712999088 0.99389374 0.0037273536 0.0023786854 1 0 2 -82 1 1 0.01018477684487269 0.9898669 0.0059374967 0.0041955533 1 0 2 -83 1 2 1.7848181575763675 0.8321596 0.16782758 1.2941472E-05 2 1 0 -84 1 1 0.40771368629306376 0.6651693 0.33338308 0.0014476714 1 2 0 -85 1 1 0.073538094082338953 0.92910075 0.061388977 0.009510201 1 2 0 -86 1 1 0.02568532133252191 0.97464174 0.023907999 0.0014502493 1 2 0 -87 1 1 0.047872087095895285 0.9532557 0.046538867 0.00020534953 1 2 0 -88 1 1 0.023007854444924543 0.9772548 0.014006073 0.008739044 1 2 0 -89 1 1 0.051250957163473519 0.9500402 0.047992557 0.0019672096 1 2 0 -90 1 1 0.088376539509711596 0.9154161 0.08403013 0.0005538561 1 2 0 -91 1 1 0.054428086534615831 0.9470266 0.051642187 0.0013313727 1 2 0 -92 1 1 0.011075027413584841 0.9889861 0.008220732 0.002793185 1 2 0 -93 1 1 0.01437425449213319 0.98572856 0.011792642 0.002478861 1 0 2 -94 1 1 0.049662849884455844 0.9515502 0.046534896 0.0019149618 1 2 0 -95 1 1 0.013598453800202021 0.9864936 0.00785744 0.0056490703 1 2 0 -96 1 1 0.025311486855273613 0.97500616 0.020882485 0.004111272 1 2 0 -97 1 1 0.01101301311502036 0.9890474 0.00809153 0.002861004 1 2 0 -98 1 1 0.081760169588047965 0.92149293 0.0778774 0.00062952057 1 0 2 -99 1 1 0.023015234494193421 0.9772476 0.018569632 0.0041829064 1 2 0 -100 2 2 0.00011158612051911557 0.9998884 0.00011122978 4.3882338E-09 2 1 0 -101 2 2 0.013729575340933843 0.98636425 0.0136345485 1.266714E-06 2 1 0 -102 2 2 0.010888574253525379 0.9891705 0.010829585 1.6294389E-07 2 1 0 -103 2 2 0.025930890575031128 0.9744024 0.025596872 6.894996E-07 2 1 0 -104 2 2 0.0015621459522632778 0.9984391 0.0015607552 3.6407382E-08 2 1 0 -105 2 2 0.0026502530714843367 0.99735326 0.0026466271 2.7263904E-09 2 1 0 -106 2 2 0.030433677359253137 0.97002476 0.0299635 1.159951E-05 2 1 0 -107 2 2 0.024480795647284091 0.9758164 0.024183862 4.4924544E-08 2 1 0 -108 2 2 0.011575682088038546 0.98849106 0.01150899 3.3041104E-08 2 1 0 -109 2 2 0.0019807750421713521 0.9980212 0.0019784786 1.6299222E-07 2 1 0 -110 2 2 0.1668877744415341 0.8462946 0.15360819 9.722446E-05 2 1 0 -111 2 2 0.031444061158942889 0.96904516 0.030953335 1.3455377E-06 2 1 0 -112 2 2 0.023395533219334613 0.976876 0.023122288 1.6099556E-06 2 1 0 -113 2 2 0.004405951750777239 0.99560374 0.0043958076 3.048161E-07 2 1 0 -114 2 2 0.00051786001087093976 0.9994823 0.00051763566 9.5351375E-08 2 1 0 -115 2 2 0.0073614401801424031 0.9926656 0.0073315674 2.597062E-06 2 1 0 -116 2 2 0.085010512269807784 0.9185026 0.08149283 4.67563E-06 2 1 0 -117 2 2 0.010475174804130676 0.9895795 0.0104205115 1.4943078E-07 2 1 0 -118 2 2 8.392686160427524E-05 0.9999161 8.4219195E-05 7.1131295E-12 2 1 0 -119 2 2 0.14243270625738888 0.8672459 0.13275184 2.228693E-06 2 1 0 -120 2 2 0.0058665359402536059 0.99415064 0.0058489824 4.4902038E-07 2 1 0 -121 2 2 0.011418315077759932 0.9886466 0.011349821 3.4921036E-06 2 1 0 -122 2 2 0.0027408576136189505 0.9972629 0.0027368125 8.6288726E-10 2 1 0 -123 2 2 0.21712063090760428 0.8048329 0.19513033 3.696039E-05 2 1 0 -124 2 2 0.018555186660661513 0.9816159 0.018382292 1.7941145E-06 2 1 0 -125 2 2 0.13317062821825568 0.8753157 0.12468223 2.215571E-06 2 1 0 -126 2 2 0.30929185172157869 0.7339665 0.26592687 0.00010682235 2 1 0 -127 2 2 0.28011468210882728 0.7556971 0.24416323 0.00013955706 2 1 0 -128 2 2 0.0030803390471882929 0.9969244 0.0030754893 6.93042E-08 2 1 0 -129 2 2 0.5607895853823287 0.5707582 0.42923495 6.927903E-06 2 1 0 -130 2 2 0.026384817227264001 0.9739602 0.026039718 7.690115E-08 2 1 0 -131 2 2 0.22129675374786231 0.8014788 0.19851315 7.966131E-06 2 1 0 -132 2 2 0.0015125382870987553 0.9984886 0.0015110222 3.6266687E-08 2 1 0 -133 2 1 0.80010965478119345 0.55066234 0.4492797 5.7917074E-05 1 2 0 -134 2 2 0.10865733918867169 0.89703774 0.10296163 7.100352E-07 2 1 0 -135 2 2 0.0060692055433251568 0.9939492 0.0060508684 4.8325873E-08 2 1 0 -136 2 2 0.0014610229435586669 0.99854004 0.0014594739 3.6343258E-07 2 1 0 -137 2 2 0.085431109138005057 0.9181164 0.08187672 6.785138E-06 2 1 0 -138 2 2 0.31214448263984934 0.7318758 0.26789987 0.00022450527 2 1 0 -139 2 2 0.059142087199894701 0.94257283 0.057418868 8.503209E-06 2 1 0 -140 2 2 0.0018873131884692992 0.99811447 0.0018852964 1.5602878E-07 2 1 0 -141 2 2 0.047074804913225708 0.95401603 0.0459597 2.4441237E-05 2 1 0 -142 2 2 0.013729575340933843 0.98636425 0.0136345485 1.266714E-06 2 1 0 -143 2 2 0.0020265834982710466 0.99797547 0.0020240443 7.197046E-08 2 1 0 -144 2 2 0.0010843193130689751 0.99891627 0.0010831588 1.3605347E-07 2 1 0 -145 2 2 0.014166328018634792 0.98593354 0.0140630165 3.5111582E-06 2 1 0 -146 2 2 0.04495238264372909 0.956043 0.04395422 2.8865643E-06 2 1 0 -147 2 2 0.067826276148074821 0.9344228 0.065563805 1.3508331E-05 2 1 0 -148 2 2 0.005008703171980048 0.9950038 0.0049937237 2.5133447E-06 2 1 0 -149 2 2 0.081482913321255249 0.92174846 0.07823061 2.0680885E-05 2 1 0 +0 0 0 0.0011100967873528965 0.9988905 0.0011098513 1.3853537E-13 0 1 2 +1 0 0 0.0070448127720684432 0.99297994 0.007020332 6.009378E-12 0 1 2 +2 0 0 0.0023012393424285556 0.9977014 0.0022987302 1.3132493E-12 0 1 2 +3 0 0 0.0071735168643098715 0.99285215 0.0071482384 1.561395E-11 0 1 2 +4 0 0 0.0007686588183116556 0.99923164 0.0007680596 9.548995E-14 0 1 2 +5 0 0 0.00070382085072599819 0.9992964 0.0007038784 1.7574012E-13 0 1 2 +6 0 0 0.0015173138878857594 0.99848384 0.0015159593 1.984035E-12 0 1 2 +7 0 0 0.0023612818220528371 0.9976415 0.0023586082 7.572865E-13 0 1 2 +8 0 0 0.010239573820286053 0.9898127 0.010187215 4.4828266E-11 0 1 2 +9 0 0 0.0076141416338779428 0.9924148 0.0075850827 3.5899447E-12 0 1 2 +10 0 0 0.00077629410864748507 0.999224 0.0007757293 3.6610746E-14 0 1 2 +11 0 0 0.0034742436669066589 0.9965318 0.0034680672 2.8527846E-12 0 1 2 +12 0 0 0.0075121045342651342 0.99251604 0.0074843164 4.123805E-12 0 1 2 +13 0 0 0.0023918122664198516 0.99761105 0.0023891912 1.5632746E-12 0 1 2 +14 0 0 8.0111851446806079E-05 0.9999199 8.048216E-05 1.7000186E-16 0 1 2 +15 0 0 5.149973917474053E-05 0.9999485 5.1264044E-05 6.434332E-16 0 1 2 +16 0 0 0.00015164571393867562 0.99984837 0.00015141307 7.743812E-15 0 1 2 +17 0 0 0.0010423726677040194 0.9989582 0.0010418568 2.6524404E-13 0 1 2 +18 0 0 0.0010814551902763913 0.9989191 0.0010809758 7.708184E-14 0 1 2 +19 0 0 0.00050545589971897188 0.9994947 0.0005050637 8.3460514E-14 0 1 2 +20 0 0 0.0050620790031108715 0.9949507 0.0050492333 1.2057628E-12 0 1 2 +21 0 0 0.00068664277443630751 0.9993136 0.00068597856 3.04746E-13 0 1 2 +22 0 0 0.00016595310050876666 0.99983406 0.00016600745 1.2559383E-14 0 1 2 +23 0 0 0.0060768214445233283 0.9939416 0.0060588485 3.6614014E-11 0 1 2 +24 0 0 0.010962512572068527 0.98909736 0.010902246 2.945331E-11 0 1 2 +25 0 0 0.015114773021998094 0.9849989 0.015001287 2.1611067E-11 0 1 2 +26 0 0 0.0030546303096060656 0.99695003 0.0030498332 6.054951E-12 0 1 2 +27 0 0 0.0016288903521015773 0.99837244 0.0016271285 2.29958E-13 0 1 2 +28 0 0 0.0016050099119122884 0.9983963 0.0016034979 2.0095434E-13 0 1 2 +29 0 0 0.0072708363197615542 0.99275553 0.0072445245 1.3592701E-11 0 1 2 +30 0 0 0.010493304889420576 0.98956156 0.0104382755 1.966337E-11 0 1 2 +31 0 0 0.0020713786723451175 0.99793077 0.00206902 9.302355E-13 0 1 2 +32 0 0 0.00018884534480816681 0.9998112 0.00018895416 2.496587E-15 0 1 2 +33 0 0 8.2973107700441223E-05 0.99991703 8.316866E-05 5.0553383E-16 0 1 2 +34 0 0 0.0076141416338779428 0.9924148 0.0075850827 3.5899447E-12 0 1 2 +35 0 0 0.0015621459522632778 0.9984391 0.0015609234 2.6512367E-13 0 1 2 +36 0 0 0.00075243401991187603 0.99924785 0.0007524573 2.1270185E-14 0 1 2 +37 0 0 0.0076141416338779428 0.9924148 0.0075850827 3.5899447E-12 0 1 2 +38 0 0 0.0048313433476961402 0.9951803 0.004819271 1.0823887E-11 0 1 2 +39 0 0 0.0023584140395519126 0.99764436 0.0023558387 5.7615056E-13 0 1 2 +40 0 0 0.00071050129341987335 0.99928975 0.0007105087 1.5976344E-13 0 1 2 +41 0 0 0.058613762568067093 0.94307095 0.056928936 1.3731134E-09 0 1 2 +42 0 0 0.0023097824802412793 0.9976929 0.0023068404 2.9820896E-12 0 1 2 +43 0 0 0.0018606199368352383 0.9981411 0.0018590544 1.1648272E-11 0 1 2 +44 0 0 0.0022039244129900587 0.9977985 0.0022013872 3.6220575E-12 0 1 2 +45 0 0 0.0066223187777712305 0.99339956 0.0066003366 1.512848E-11 0 1 2 +46 0 0 0.00079061043515947244 0.9992097 0.00078995415 9.513509E-14 0 1 2 +47 0 0 0.0033817186504715878 0.996624 0.0033761603 3.763931E-12 0 1 2 +48 0 0 0.00077724852403853091 0.99922305 0.00077664276 4.8120887E-14 0 1 2 +49 0 0 0.0023269884618295112 0.9976757 0.0023243777 6.6178313E-13 0 1 2 +50 1 1 0.0060296277781084866 0.9939885 0.0040062875 0.0020049622 1 2 0 +51 1 1 0.023376923644402699 0.9768942 0.018610137 0.0044954917 1 2 0 +52 1 1 0.031613839401899838 0.96888065 0.030452834 0.0006662728 1 2 0 +53 1 1 0.085075602344142798 0.91844285 0.08065317 0.0009038801 1 2 0 +54 1 1 0.063746903397246202 0.93824244 0.061086345 0.0006710334 1 2 0 +55 1 1 0.089693777542918976 0.9142111 0.084950805 0.00083812367 1 2 0 +56 1 1 0.085069177512258412 0.91844875 0.07853346 0.0030178856 1 2 0 +57 1 1 0.019546522245977516 0.9806433 0.016897747 0.0024591105 1 0 2 +58 1 1 0.0099081882770904756 0.99014074 0.008955239 0.0009040547 1 2 0 +59 1 1 0.092630883817733156 0.9115299 0.0823354 0.006134713 1 2 0 +60 1 1 0.014457703268027379 0.9856463 0.012556279 0.0017973994 1 2 0 +61 1 1 0.045186017212082943 0.95581967 0.03756768 0.006612597 1 2 0 +62 1 1 0.0040554869713392292 0.9959527 0.0034838247 0.00056346314 1 2 0 +63 1 1 0.10219478993047783 0.90285367 0.09655194 0.0005945007 1 2 0 +64 1 1 0.043745547946894225 0.9571975 0.0403017 0.002500794 1 0 2 +65 1 1 0.0080270808006034106 0.99200505 0.0043637045 0.003631049 1 0 2 +66 1 1 0.25794669864881997 0.7726364 0.22568193 0.0016817307 1 2 0 +67 1 1 0.004681082317533663 0.99532986 0.0024284304 0.0022418906 1 0 2 +68 1 1 0.4186439374882594 0.6579384 0.3419866 7.495434E-05 1 2 0 +69 1 1 0.0088572595356648746 0.99118185 0.006171344 0.002646689 1 2 0 +70 1 2 1.1223549691918893 0.67391926 0.32551232 0.0005684842 2 1 0 +71 1 1 0.01054667309481289 0.98950875 0.006225767 0.0042653275 1 0 2 +72 1 1 0.6119784875601737 0.5422769 0.45768276 4.030501E-05 1 2 0 +73 1 1 0.033711439152709112 0.96685046 0.032761887 0.00038760784 1 2 0 +74 1 1 0.0076026701985775378 0.99242616 0.004710559 0.002863219 1 2 0 +75 1 1 0.0093264821417960372 0.9907169 0.0062752664 0.003007812 1 2 0 +76 1 1 0.030996808891834982 0.96947867 0.030218327 0.00030285056 1 2 0 +77 1 1 0.3659636701213016 0.693528 0.30621776 0.00025408115 1 2 0 +78 1 1 0.12320159568004112 0.8840854 0.114578724 0.0013359514 1 2 0 +79 1 1 0.017106038862047986 0.98303944 0.016605409 0.00035523012 1 0 2 +80 1 1 0.0099058405518300738 0.99014306 0.007177006 0.0026797957 1 2 0 +81 1 1 0.0061030277839462773 0.99391556 0.003708176 0.0023763352 1 0 2 +82 1 1 0.010160932065713054 0.9898905 0.0059115994 0.0041978564 1 0 2 +83 1 2 1.7840146780311217 0.83202463 0.16796248 1.2869889E-05 2 1 0 +84 1 1 0.40920238010603066 0.6641798 0.33437788 0.0014422628 1 2 0 +85 1 1 0.073859680860476701 0.928802 0.06170616 0.0094919205 1 2 0 +86 1 1 0.025666241017175091 0.97466034 0.023897348 0.0014421142 1 2 0 +87 1 1 0.047637449122048356 0.9534794 0.046317182 0.00020365046 1 2 0 +88 1 1 0.023036216088822219 0.9772271 0.014055676 0.008717236 1 2 0 +89 1 1 0.051268336038284672 0.9500237 0.04801822 0.0019581784 1 2 0 +90 1 1 0.088363647401615383 0.9154279 0.08402101 0.00055107626 1 2 0 +91 1 1 0.054470508133182577 0.94698644 0.051688343 0.0013252622 1 2 0 +92 1 1 0.011057368917507868 0.98900354 0.008217381 0.0027790256 1 2 0 +93 1 1 0.014335132717320211 0.9857671 0.011750958 0.0024818496 1 0 2 +94 1 1 0.049708577777940119 0.9515067 0.046586514 0.0019068395 1 2 0 +95 1 1 0.013599843477539789 0.9864922 0.007876568 0.005631309 1 2 0 +96 1 1 0.025343398572549754 0.97497505 0.020928022 0.0040969974 1 2 0 +97 1 1 0.010998730483251765 0.98906153 0.008091495 0.0028467982 1 2 0 +98 1 1 0.08160397308321414 0.9216369 0.077731304 0.0006317606 1 0 2 +99 1 1 0.0230314585956332 0.97723174 0.018601017 0.0041670897 1 2 0 +100 2 2 0.00011063234023480313 0.9998894 0.00011070922 4.351013E-09 2 1 0 +101 2 2 0.013711446914811879 0.9863821 0.013616532 1.2585332E-06 2 1 0 +102 2 2 0.01091575062028363 0.9891436 0.010856191 1.6200677E-07 2 1 0 +103 2 2 0.025952361637010559 0.9743815 0.025617588 6.854851E-07 2 1 0 +104 2 2 0.0015611907874603366 0.99844 0.0015597419 3.6159353E-08 2 1 0 +105 2 2 0.0026664489273118116 0.9973371 0.00266266 2.7131466E-09 2 1 0 +106 2 2 0.030318717531012056 0.9701363 0.02985217 1.1521343E-05 2 1 0 +107 2 2 0.024636260966751176 0.97566473 0.02433561 4.4725848E-08 2 1 0 +108 2 2 0.01164243488614073 0.9884251 0.011575128 3.289459E-08 2 1 0 +109 2 2 0.0019750416673904813 0.9980269 0.0019724772 1.617013E-07 2 1 0 +110 2 2 0.16647675820074742 0.8466425 0.15326092 9.659204E-05 2 1 0 +111 2 2 0.031490808813879896 0.96899986 0.030998776 1.3378963E-06 2 1 0 +112 2 2 0.023404136451361482 0.9768676 0.023130763 1.5998768E-06 2 1 0 +113 2 2 0.0044012222026529255 0.99560845 0.004391366 3.0283564E-07 2 1 0 +114 2 2 0.00051595167606361596 0.9994842 0.0005152728 9.4579214E-08 2 1 0 +115 2 2 0.0073308777219288574 0.9926959 0.007301546 2.5768245E-06 2 1 0 +116 2 2 0.085065867766488085 0.9184518 0.081543766 4.6483183E-06 2 1 0 +117 2 2 0.010478969445945856 0.98957574 0.010424239 1.4843683E-07 2 1 0 +118 2 2 8.4880616417756642E-05 0.9999151 8.498341E-05 7.0823174E-12 2 1 0 +119 2 2 0.14308164654434841 0.8666833 0.13331467 2.218199E-06 2 1 0 +120 2 2 0.0058594012796195543 0.99415773 0.0058418275 4.458627E-07 2 1 0 +121 2 2 0.011374968132936548 0.9886895 0.011307055 3.4667921E-06 2 1 0 +122 2 2 0.0027647054244224279 0.9972391 0.0027604867 8.594396E-10 2 1 0 +123 2 2 0.21725987041570646 0.8047208 0.19524235 3.6747544E-05 2 1 0 +124 2 2 0.018522762201905772 0.98164773 0.018350681 1.7817254E-06 2 1 0 +125 2 2 0.1335644992952405 0.87497103 0.12502682 2.2033246E-06 2 1 0 +126 2 2 0.30912157107031513 0.7340915 0.26580217 0.00010620019 2 1 0 +127 2 2 0.27955964400421202 0.7561166 0.24374475 0.00013872533 2 1 0 +128 2 2 0.0030812956641316051 0.99692345 0.0030768916 6.886699E-08 2 1 0 +129 2 2 0.56300973710059499 0.56949246 0.4305007 6.883423E-06 2 1 0 +130 2 2 0.026556982677033589 0.97379255 0.026207458 7.655899E-08 2 1 0 +131 2 2 0.22159337865988013 0.8012411 0.19875106 7.916578E-06 2 1 0 +132 2 2 0.0015125382870987553 0.9984886 0.0015109567 3.6027792E-08 2 1 0 +133 2 1 0.80118038716036721 0.5511436 0.4487989 5.757113E-05 1 2 0 +134 2 2 0.10909272224640143 0.8966473 0.10335219 7.0668176E-07 2 1 0 +135 2 2 0.0060958914358296692 0.99392265 0.0060769496 4.806114E-08 2 1 0 +136 2 2 0.0014524273624274906 0.9985486 0.0014510206 3.6032796E-07 2 1 0 +137 2 2 0.085394364767160641 0.9181501 0.08184301 6.7438123E-06 2 1 0 +138 2 2 0.31136751231271187 0.73244464 0.26733193 0.00022317962 2 1 0 +139 2 2 0.05913209594350867 0.94258225 0.05740952 8.44917E-06 2 1 0 +140 2 2 0.0018834912921805113 0.9981183 0.0018814779 1.5487406E-07 2 1 0 +141 2 2 0.046989901440894453 0.95409703 0.04587865 2.427137E-05 2 1 0 +142 2 2 0.013711446914811879 0.9863821 0.013616532 1.2585332E-06 2 1 0 +143 2 2 0.0020237166754501316 0.9979783 0.002021656 7.145971E-08 2 1 0 +144 2 2 0.0010795457796386584 0.99892104 0.001079097 1.3495621E-07 2 1 0 +145 2 2 0.014142509019330281 0.985957 0.014039472 3.4864931E-06 2 1 0 +146 2 2 0.045035866284553396 0.9559632 0.044033695 2.8706036E-06 2 1 0 +147 2 2 0.067760959715770355 0.9344838 0.06550267 1.3421985E-05 2 1 0 +148 2 2 0.0049781526314697216 0.9950342 0.0049631833 2.4921453E-06 2 1 0 +149 2 2 0.081277817729632454 0.9219375 0.07804192 2.0546579E-05 2 1 0 diff --git a/test/BaselineOutput/Common/SymSGD/netcoreapp/SymSGD-CV-breast-cancer.txt b/test/BaselineOutput/Common/SymSGD/netcoreapp/SymSGD-CV-breast-cancer.txt new file mode 100644 index 0000000000..d68782c62c --- /dev/null +++ b/test/BaselineOutput/Common/SymSGD/netcoreapp/SymSGD-CV-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +5 1 1000.1891 1 -0 1 +6 0 95.69922 1 Infinity 1 +8 0 -334.17828 0 -0 0 +9 0 -235.6391 0 -0 0 +10 0 -277.29367 0 -0 0 +11 0 -322.8632 0 -0 0 +18 1 646.0083 1 -0 1 +20 1 198.0318 1 -0 1 +21 1 459.70312 1 -0 1 +25 1 124.40161 1 -0 1 +28 0 -322.8632 0 -0 0 +31 0 -300.14307 0 -0 0 +32 1 450.40417 1 -0 1 +35 0 -322.8632 0 -0 0 +37 0 -82.94397 9.503676E-37 -0 0 +40 0 ? ? ? 0 +41 1 193.508 1 -0 1 +44 1 640.1284 1 -0 1 +45 0 -325.427 0 -0 0 +46 1 662.572 1 -0 1 +48 0 -313.28973 0 -0 0 +50 1 277.65234 1 -0 1 +51 1 44.778015 1 -0 1 +52 1 271.5711 1 -0 1 +54 1 311.60034 1 -0 1 +56 1 904.77344 1 -0 1 +60 1 94.159424 1 -0 1 +63 1 -9.789154 5.6053155E-05 14.12284488867504 0 +64 0 -328.27826 0 -0 0 +66 0 -289.31293 0 -0 0 +68 1 540.3833 1 -0 1 +69 0 -294.4256 0 -0 0 +70 0 -261.91486 0 -0 0 +71 1 360.4336 1 -0 1 +72 0 -18.042511 1.45961065E-08 2.105773068336081E-08 0 +73 1 337.30237 1 -0 1 +74 1 314.5249 1 -0 1 +76 0 -261.22357 0 -0 0 +77 0 -165.81848 0 -0 0 +79 0 -351.12762 0 -0 0 +82 0 -207.50394 0 -0 0 +88 0 -289.31293 0 -0 0 +90 0 -305.55814 0 -0 0 +91 0 -352.15662 0 -0 0 +92 0 -289.31293 0 -0 0 +93 0 -328.27826 0 -0 0 +95 0 -305.55814 0 -0 0 +96 0 -357.5717 0 -0 0 +97 0 -283.8979 0 -0 0 +98 1 375.3628 1 -0 1 +99 1 541.3288 1 -0 1 +100 1 40.636658 1 -0 1 +102 0 -266.9614 0 -0 0 +104 1 509.3629 1 -0 1 +105 1 -0.046447754 0.48839015 1.0338939978132953 0 +106 1 971.9457 1 -0 1 +108 0 -306.01318 0 -0 0 +109 1 561.1019 1 -0 1 +111 1 429.27478 1 -0 1 +112 1 352.87836 1 -0 1 +113 1 807.48413 1 -0 1 +115 0 -172.66177 0 -0 0 +117 1 497.8714 1 -0 1 +120 0 -295.19498 0 -0 0 +121 0 -168.33484 0 -0 0 +122 1 868.4481 1 -0 1 +123 1 257.67218 1 -0 1 +125 0 -328.27826 0 -0 0 +128 1 268.114 1 -0 1 +129 0 -577.114 0 -0 0 +131 0 -300.14307 0 -0 0 +132 1 745.2488 1 -0 1 +133 0 -306.31558 0 -0 0 +137 0 -340.16824 0 -0 0 +138 0 -289.68152 0 -0 0 +141 0 -345.5833 0 -0 0 +144 0 -322.8632 0 -0 0 +145 0 ? ? ? 0 +147 0 -311.95856 0 -0 0 +150 0 -277.29367 0 -0 0 +151 1 243.12073 1 -0 1 +152 1 862.843 1 -0 1 +154 0 -350.99835 0 -0 0 +156 0 -229.53273 0 -0 0 +161 0 -278.1804 0 -0 0 +164 0 ? ? ? 0 +167 1 283.807 1 -0 1 +169 0 -334.7223 0 -0 0 +171 0 -305.55814 0 -0 0 +173 1 1010.343 1 -0 1 +174 1 608.9114 1 -0 1 +176 0 -300.14307 0 -0 0 +177 1 562.58374 1 -0 1 +179 1 177.81586 1 -0 1 +180 0 -277.29367 0 -0 0 +181 0 -350.99835 0 -0 0 +183 1 808.6741 1 -0 1 +187 1 524.15204 1 -0 1 +188 1 673.8281 1 -0 1 +189 0 -184.46106 0 -0 0 +191 1 429.5337 1 -0 1 +192 0 -306.61798 0 -0 0 +196 0 363.26575 1 Infinity 1 +198 0 -350.99835 0 -0 0 +199 0 -317.44812 0 -0 0 +201 1 687.1731 1 -0 1 +202 0 -305.55814 0 -0 0 +204 0 -305.55814 0 -0 0 +205 1 999.5823 1 -0 1 +206 1 667.10315 1 -0 1 +207 0 -277.29367 0 -0 0 +209 0 -255.63342 0 -0 0 +210 1 1057.675 1 -0 1 +211 1 905.6708 1 -0 1 +212 0 -305.55814 0 -0 0 +216 0 -328.27826 0 -0 0 +218 1 708.58105 1 -0 1 +219 0 -233.7797 0 -0 0 +223 1 528.3917 1 -0 1 +226 1 694.28674 1 -0 1 +228 0 -277.29367 0 -0 0 +233 1 352.99603 1 -0 1 +237 1 444.02045 1 -0 1 +239 1 448.62347 1 -0 1 +240 0 -219.04243 0 -0 0 +241 0 -290.44034 0 -0 0 +242 0 -300.14307 0 -0 0 +244 0 -305.55814 0 -0 0 +246 1 1022.4103 1 -0 1 +247 1 407.5221 1 -0 1 +248 0 -233.65024 0 -0 0 +249 0 ? ? ? 0 +250 0 -252.25284 0 -0 0 +252 0 294.28015 1 Infinity 1 +254 1 734.71045 1 -0 1 +257 0 -317.44812 0 -0 0 +258 0 -294.728 0 -0 0 +259 0 569.26843 1 Infinity 1 +260 1 521.4542 1 -0 1 +262 1 895.22437 1 -0 1 +267 1 399.35107 1 -0 1 +268 1 728.25024 1 -0 1 +269 0 -305.55814 0 -0 0 +271 0 -283.8979 0 -0 0 +272 1 399.35107 1 -0 1 +275 0 ? ? ? 0 +276 0 -317.44812 0 -0 0 +277 0 -328.27826 0 -0 0 +278 0 -305.55814 0 -0 0 +279 1 440.3501 1 -0 1 +280 0 -294.728 0 -0 0 +283 1 553.0797 1 -0 1 +284 1 419.6612 1 -0 1 +285 1 976.37036 1 -0 1 +288 1 33.66107 1 -0 1 +290 0 -350.99835 0 -0 0 +291 0 -305.55814 0 -0 0 +293 1 380.2743 1 -0 1 +296 0 123.29089 1 Infinity 1 +297 0 ? ? ? 0 +299 1 205.4273 1 -0 1 +300 1 383.03644 1 -0 1 +301 0 -305.55814 0 -0 0 +303 0 -305.55814 0 -0 0 +304 1 251.54071 1 -0 1 +308 1 568.57874 1 -0 1 +309 0 -70.95117 1.5356689E-31 -0 0 +311 0 -350.99835 0 -0 0 +312 1 -108.10928 0 Infinity 0 +314 0 -300.0138 0 -0 0 +316 1 460.54492 1 -0 1 +317 1 708.1575 1 -0 1 +319 0 149.38269 1 Infinity 1 +321 0 ? ? ? 0 +323 1 373.12366 1 -0 1 +327 0 -328.27826 0 -0 0 +328 1 573.4951 1 -0 1 +329 1 455.49634 1 -0 1 +331 0 -281.04663 0 -0 0 +332 0 -208.72833 0 -0 0 +333 1 386.81335 1 -0 1 +336 1 410.70654 1 -0 1 +338 0 -300.0138 0 -0 0 +343 0 -350.99835 0 -0 0 +344 1 453.73303 1 -0 1 +346 0 -248.58038 0 -0 0 +347 0 -300.46884 0 -0 0 +348 1 -168.91534 0 Infinity 0 +349 1 214.19824 1 -0 1 +350 0 -353.31488 0 -0 0 +352 0 26.112091 1 Infinity 1 +353 1 631.7583 1 -0 1 +354 0 -328.27826 0 -0 0 +355 0 -329.16498 0 -0 0 +358 1 578.1035 1 -0 1 +360 1 916.9276 1 -0 1 +361 1 671.8186 1 -0 1 +366 1 942.1826 1 -0 1 +368 0 -306.29526 0 -0 0 +370 0 -165.15805 0 -0 0 +371 0 -306.29526 0 -0 0 +373 0 -318.60638 0 -0 0 +376 0 -328.27826 0 -0 0 +377 0 -300.0138 0 -0 0 +378 0 -370.54523 0 -0 0 +379 0 -120.58423 0 -0 0 +381 1 631.9988 1 -0 1 +383 0 -345.5833 0 -0 0 +384 0 -345.5833 0 -0 0 +387 0 -126.6597 0 -0 0 +388 0 -311.73065 0 -0 0 +389 0 -279.16577 0 -0 0 +391 1 929.01294 1 -0 1 +392 0 -317.44812 0 -0 0 +395 0 -317.44812 0 -0 0 +396 0 -294.728 0 -0 0 +398 0 -229.92163 0 -0 0 +399 0 -229.9013 0 -0 0 +404 0 -287.44864 0 -0 0 +406 0 -213.28755 0 -0 0 +409 0 -295.09656 0 -0 0 +413 0 -261.54633 0 -0 0 +414 1 634.50854 1 -0 1 +415 0 -74.0965 6.611506E-33 -0 0 +416 1 576.37305 1 -0 1 +418 0 -138.57281 0 -0 0 +419 0 -277.58878 0 -0 0 +422 0 -69.55682 6.1923467E-31 -0 0 +423 0 -261.91486 0 -0 0 +428 0 -328.27826 0 -0 0 +429 0 -322.8632 0 -0 0 +430 0 -164.35779 0 -0 0 +434 0 679.9729 1 Infinity 1 +436 1 574.03687 1 -0 1 +439 0 -329.3381 0 -0 0 +440 1 412.9361 1 -0 1 +441 0 -132.80533 0 -0 0 +442 0 -285.80685 0 -0 0 +449 1 771.35254 1 -0 1 +450 0 -307.74536 0 -0 0 +451 0 -329.3381 0 -0 0 +452 0 -364.14502 0 -0 0 +453 1 535.7859 1 -0 1 +454 0 -221.05986 0 -0 0 +455 1 22.2547 1 -0 1 +456 1 616.36194 1 -0 1 +457 1 621.3196 1 -0 1 +464 0 -334.75317 0 -0 0 +465 1 656.32996 1 -0 1 +466 1 772.76086 1 -0 1 +467 1 661.41406 1 -0 1 +474 0 -329.3381 0 -0 0 +480 0 -301.07364 0 -0 0 +482 1 755.6924 1 -0 1 +483 1 852.1958 1 -0 1 +484 0 -306.98654 0 -0 0 +487 1 937.19653 1 -0 1 +489 1 24.864746 1 -0 1 +492 0 -283.7686 0 -0 0 +493 1 900.6575 1 -0 1 +495 0 -289.18365 0 -0 0 +497 0 -258.35922 0 -0 0 +501 0 -312.03305 0 -0 0 +502 0 -324.02142 0 -0 0 +504 0 -350.99835 0 -0 0 +507 0 -217.24591 0 -0 0 +510 0 -350.99835 0 -0 0 +513 0 -289.18365 0 -0 0 +514 1 719.33826 1 -0 1 +517 0 -300.0138 0 -0 0 +519 1 782.5841 1 -0 1 +520 0 -379.26282 0 -0 0 +521 0 -363.01764 0 -0 0 +522 1 289.018 1 -0 1 +523 1 448.80402 1 -0 1 +527 0 -289.31293 0 -0 0 +528 0 -291.2083 0 -0 0 +529 0 -283.7686 0 -0 0 +531 0 -213.28755 0 -0 0 +532 0 -277.29367 0 -0 0 +533 0 -317.44812 0 -0 0 +534 0 -322.8632 0 -0 0 +535 0 -272.8434 0 -0 0 +538 0 -312.03305 0 -0 0 +539 0 -301.20294 0 -0 0 +540 0 -262.03384 0 -0 0 +541 0 -340.16824 0 -0 0 +544 0 -290.14847 0 -0 0 +546 1 1038.1329 1 -0 1 +547 0 -317.31885 0 -0 0 +548 0 -311.90378 0 -0 0 +549 1 527.9298 1 -0 1 +557 0 -353.31488 0 -0 0 +558 0 -322.8632 0 -0 0 +559 0 -306.61798 0 -0 0 +560 0 -283.8979 0 -0 0 +561 0 -283.8979 0 -0 0 +563 0 -317.44812 0 -0 0 +565 1 843.4319 1 -0 1 +566 0 -272.37646 0 -0 0 +569 1 725.87305 1 -0 1 +577 0 -328.27826 0 -0 0 +578 0 -328.27826 0 -0 0 +581 1 767.5625 1 -0 1 +582 1 963.59705 1 -0 1 +584 0 -417.54453 0 -0 0 +586 1 1060.5945 1 -0 1 +590 1 609.25183 1 -0 1 +593 0 -306.98654 0 -0 0 +594 1 768.8346 1 -0 1 +600 0 -317.44812 0 -0 0 +602 0 -312.03305 0 -0 0 +604 1 240.23865 1 -0 1 +606 0 -346.74155 0 -0 0 +607 0 -350.99835 0 -0 0 +609 0 -329.3381 0 -0 0 +612 1 1086.2559 1 -0 1 +613 0 -170.07529 0 -0 0 +614 0 -294.59872 0 -0 0 +617 0 ? ? ? 0 +618 0 -312.03305 0 -0 0 +619 0 -306.61798 0 -0 0 +621 0 -20.488556 1.2645419E-09 1.8243483518208995E-09 0 +622 0 -297.4131 0 -0 0 +624 0 -290.34192 0 -0 0 +627 0 -167.2467 0 -0 0 +629 0 -334.75317 0 -0 0 +633 1 371.80853 1 -0 1 +634 0 -340.16824 0 -0 0 +638 0 -334.75317 0 -0 0 +639 0 -353.31488 0 -0 0 +641 0 -317.44812 0 -0 0 +642 0 -317.44812 0 -0 0 +644 0 -345.5833 0 -0 0 +645 0 -317.44812 0 -0 0 +649 0 -317.44812 0 -0 0 +652 0 -295.48544 0 -0 0 +653 0 -312.03305 0 -0 0 +654 0 -294.728 0 -0 0 +656 0 -306.61798 0 -0 0 +657 0 -79.233826 3.883173E-35 -0 0 +660 0 -328.27826 0 -0 0 +661 0 -289.31293 0 -0 0 +665 0 -350.99835 0 -0 0 +668 1 325.23853 1 -0 1 +670 1 794.5818 1 -0 1 +678 0 -350.99835 0 -0 0 +679 0 -345.5833 0 -0 0 +680 1 1116.7966 1 -0 1 +681 1 948.44995 1 -0 1 +682 0 -272.76532 0 -0 0 +683 0 -350.99835 0 -0 0 +685 0 -350.99835 0 -0 0 +688 0 -334.75317 0 -0 0 +689 0 -331.03958 0 -0 0 +691 1 734.78345 1 -0 1 +692 0 -340.16824 0 -0 0 +693 0 -314.59686 0 -0 0 +694 0 -323.84854 0 -0 0 +696 1 745.32556 1 -0 1 +697 1 641.42883 1 -0 1 +698 1 668.19824 1 -0 1 +0 0 -655.84644 0 -0 0 +1 0 352.09045 1 Infinity 1 +2 0 -559.3663 0 -0 0 +3 0 1740.6991 1 Infinity 1 +4 0 -568.0045 0 -0 0 +7 0 -495.54675 0 -0 0 +12 1 241.31317 1 -0 1 +13 0 -462.8861 0 -0 0 +14 1 942.8898 1 -0 1 +15 1 16.798767 0.99999994 8.5991327994145617E-08 1 +16 0 -553.6886 0 -0 0 +17 0 -643.05927 0 -0 0 +19 0 -668.63354 0 -0 0 +22 0 -540.9015 0 -0 0 +23 1 ? ? ? 0 +24 0 -604.6978 0 -0 0 +26 0 -270.65833 0 -0 0 +27 0 -566.47577 0 -0 0 +29 0 -182.08011 0 -0 0 +30 0 -411.08737 0 -0 0 +33 0 -579.9576 0 -0 0 +34 0 -418.96326 0 -0 0 +36 1 1591.2499 1 -0 1 +38 1 1299.1012 1 -0 1 +39 1 138.43103 1 -0 1 +42 1 1330.6154 1 -0 1 +43 1 -316.8241 0 Infinity 0 +47 0 -515.32715 0 -0 0 +49 1 1846.6259 1 -0 1 +53 1 -274.02063 0 Infinity 0 +55 1 1099.3195 1 -0 1 +57 1 -608.20197 0 Infinity 0 +58 1 -331.38577 0 Infinity 0 +59 1 272.1659 1 -0 1 +61 0 -444.42126 0 -0 0 +62 1 1275.4248 1 -0 1 +65 1 -452.49982 0 Infinity 0 +67 1 413.8687 1 -0 1 +75 0 -419.58133 0 -0 0 +78 0 -488.46042 0 -0 0 +80 0 -582.7864 0 -0 0 +81 0 -516.16125 0 -0 0 +83 0 -916.849 0 -0 0 +84 1 897.4657 1 -0 1 +85 1 746.3778 1 -0 1 +86 1 635.3164 1 -0 1 +87 1 1356.6456 1 -0 1 +89 0 -620.40094 0 -0 0 +94 0 -617.485 0 -0 0 +101 1 841.3364 1 -0 1 +103 1 -1044.8297 0 Infinity 0 +107 1 1459.1467 1 -0 1 +110 0 -263.18948 0 -0 0 +114 0 -85.66412 6.2595604E-38 -0 0 +116 0 -16.924438 4.4648793E-08 6.4414593835410614E-08 0 +118 0 -543.7728 0 -0 0 +119 0 -418.93723 0 -0 0 +124 1 404.9671 1 -0 1 +126 1 567.6393 1 -0 1 +127 0 -630.2721 0 -0 0 +130 0 -322.59937 0 -0 0 +134 0 -670.71545 0 -0 0 +135 0 -421.6544 0 -0 0 +136 0 -553.6886 0 -0 0 +139 0 ? ? ? 0 +140 0 -451.53076 0 -0 0 +142 1 488.31128 1 -0 1 +143 0 -142.33295 0 -0 0 +146 1 204.00928 1 -0 1 +148 0 -941.34155 0 -0 0 +149 1 752.699 1 -0 1 +153 0 -322.50616 0 -0 0 +155 1 1089.5981 1 -0 1 +157 0 -617.485 0 -0 0 +158 0 ? ? ? 0 +159 1 1923.4984 1 -0 1 +160 1 1494.5034 1 -0 1 +162 0 -630.2721 0 -0 0 +163 0 -310.77014 0 -0 0 +165 0 -490.5103 0 -0 0 +166 1 1570.0204 1 -0 1 +168 0 -630.2721 0 -0 0 +170 0 -451.53076 0 -0 0 +172 0 -515.32715 0 -0 0 +175 1 1194.7078 1 -0 1 +178 0 -643.05927 0 -0 0 +182 0 -668.63354 0 -0 0 +184 1 1070.4229 1 -0 1 +185 0 -487.67093 0 -0 0 +186 1 1482.4264 1 -0 1 +190 1 2258.6045 1 -0 1 +193 0 -604.6978 0 -0 0 +194 0 -630.2721 0 -0 0 +195 0 -643.05927 0 -0 0 +197 0 -543.2643 0 -0 0 +200 1 1415.3917 1 -0 1 +203 0 -655.84644 0 -0 0 +208 0 -474.88376 0 -0 0 +213 1 2248.6738 1 -0 1 +214 1 2270.229 1 -0 1 +215 1 1644.9979 1 -0 1 +217 0 -604.6978 0 -0 0 +220 0 -567.1704 0 -0 0 +221 1 221.4364 1 -0 1 +222 1 -65.19629 4.848545E-29 94.058362912090217 0 +224 1 1288.9967 1 -0 1 +225 0 -515.32715 0 -0 0 +227 1 1841.0248 1 -0 1 +229 1 1514.782 1 -0 1 +230 1 930.4315 1 -0 1 +231 1 1347.1566 1 -0 1 +232 0 355.9265 1 Infinity 1 +234 0 50.925537 1 Infinity 1 +235 0 ? ? ? 0 +236 1 1204.765 1 -0 1 +238 1 2421.6172 1 -0 1 +243 0 -499.81567 0 -0 0 +245 0 -547.4132 0 -0 0 +251 1 963.10803 1 -0 1 +253 1 1330.6154 1 -0 1 +255 1 1480.2573 1 -0 1 +256 0 -451.53076 0 -0 0 +261 1 1146.0453 1 -0 1 +263 1 671.7936 1 -0 1 +264 1 168.76099 1 -0 1 +265 0 -208.3888 0 -0 0 +266 1 1778.7516 1 -0 1 +270 1 1587.8641 1 -0 1 +273 1 71.01257 1 -0 1 +274 0 -548.629 0 -0 0 +281 0 -579.9576 0 -0 0 +282 1 1006.2108 1 -0 1 +286 1 1933.3766 1 -0 1 +287 0 -670.71545 0 -0 0 +289 1 1586.0116 1 -0 1 +292 1 ? ? ? 0 +294 0 ? ? ? 0 +295 1 906.5337 1 -0 1 +298 0 -764.4797 0 -0 0 +302 1 1682.4965 1 -0 1 +305 1 1757.9471 1 -0 1 +306 0 -604.6978 0 -0 0 +307 0 -604.6978 0 -0 0 +310 0 -657.92834 0 -0 0 +313 0 -425.95645 0 -0 0 +315 0 ? ? ? 0 +318 0 -994.1401 0 -0 0 +320 1 276.64636 1 -0 1 +322 0 -630.2721 0 -0 0 +324 0 -604.6978 0 -0 0 +325 0 -115.24805 0 -0 0 +326 1 -15.441406 1.9673534E-07 22.27724050916763 0 +330 1 698.8905 1 -0 1 +334 1 1535.5616 1 -0 1 +335 0 -425.95645 0 -0 0 +337 0 -604.6978 0 -0 0 +339 1 1217.4391 1 -0 1 +340 1 493.83777 1 -0 1 +341 0 -604.6978 0 -0 0 +342 0 -438.74362 0 -0 0 +345 0 -425.95645 0 -0 0 +351 0 -617.485 0 -0 0 +356 1 -20.483704 1.2706928E-09 29.551737594786047 0 +357 1 1071.4106 1 -0 1 +359 1 718.9486 1 -0 1 +362 0 -396.35065 0 -0 0 +363 0 412.75226 1 Infinity 1 +364 0 -617.485 0 -0 0 +365 0 -528.1143 0 -0 0 +367 1 1990.4601 1 -0 1 +369 0 -141.63672 0 -0 0 +372 0 -431.7504 0 -0 0 +374 0 -418.96326 0 -0 0 +375 0 -425.95645 0 -0 0 +380 0 -425.95645 0 -0 0 +382 0 -248.73444 0 -0 0 +385 0 -124.03369 0 -0 0 +386 1 986.1404 1 -0 1 +390 0 -477.79977 0 -0 0 +393 0 -296.1424 0 -0 0 +394 0 -131.02173 0 -0 0 +397 0 -464.3179 0 -0 0 +400 1 1161.0793 1 -0 1 +401 0 -451.53076 0 -0 0 +402 0 -41.741272 7.44725E-19 -0 0 +403 0 -238.81316 0 -0 0 +405 0 -515.32715 0 -0 0 +407 0 -515.32715 0 -0 0 +408 0 -106.25546 0 -0 0 +410 0 -515.32715 0 -0 0 +411 0 ? ? ? 0 +412 1 1142.7122 1 -0 1 +417 0 -515.32715 0 -0 0 +420 0 -151.03839 0 -0 0 +421 1 1101.1173 1 -0 1 +424 0 -451.53076 0 -0 0 +425 1 1700.8898 1 -0 1 +426 0 413.44397 1 Infinity 1 +427 1 1406.7853 1 -0 1 +431 0 -758.77673 0 -0 0 +432 0 -484.83264 0 -0 0 +433 0 -114.108734 0 -0 0 +435 1 1690.2623 1 -0 1 +437 0 -464.3179 0 -0 0 +438 0 -145.38675 0 -0 0 +443 0 -355.0506 0 -0 0 +444 0 -508.6529 0 -0 0 +445 0 -438.74362 0 -0 0 +446 0 -425.95645 0 -0 0 +447 0 -477.10504 0 -0 0 +448 0 -296.1424 0 -0 0 +458 0 -355.16684 0 -0 0 +459 0 -233.22867 0 -0 0 +460 0 -402.05032 0 -0 0 +461 0 -167.90726 0 -0 0 +462 0 -414.83746 0 -0 0 +463 0 -382.6748 0 -0 0 +468 0 -464.3179 0 -0 0 +469 0 -393.38895 0 -0 0 +470 0 -411.08737 0 -0 0 +471 0 -414.83746 0 -0 0 +472 0 -360.07812 0 -0 0 +473 0 -464.3179 0 -0 0 +475 0 -451.53076 0 -0 0 +476 0 -342.3797 0 -0 0 +477 0 -464.3179 0 -0 0 +478 0 -336.676 0 -0 0 +479 1 1756.934 1 -0 1 +481 0 38.273315 1 Infinity 1 +485 0 -79.27588 3.723259E-35 -0 0 +486 0 -411.08737 0 -0 0 +488 1 1032.361 1 -0 1 +490 0 -425.95645 0 -0 0 +491 1 1566.1005 1 -0 1 +494 0 -82.79584 1.1021094E-36 -0 0 +496 0 -296.1424 0 -0 0 +498 0 -553.6886 0 -0 0 +499 0 -553.6886 0 -0 0 +500 0 -668.63354 0 -0 0 +503 0 -643.05927 0 -0 0 +505 0 -170.6727 0 -0 0 +506 1 1927.5609 1 -0 1 +508 0 -477.10504 0 -0 0 +509 0 -438.74362 0 -0 0 +511 0 -566.47577 0 -0 0 +512 0 -477.10504 0 -0 0 +515 1 1918.6837 1 -0 1 +516 0 -296.1424 0 -0 0 +518 0 -292.0652 0 -0 0 +524 0 -540.9015 0 -0 0 +525 0 -414.0034 0 -0 0 +526 0 -464.3179 0 -0 0 +530 1 944.2931 1 -0 1 +536 0 -655.84644 0 -0 0 +537 0 -533.9082 0 -0 0 +542 0 -196.24692 0 -0 0 +543 0 -553.6886 0 -0 0 +545 0 -566.47577 0 -0 0 +550 0 -540.9015 0 -0 0 +551 0 -604.6978 0 -0 0 +552 0 -338.1051 0 -0 0 +553 0 240.83246 1 Infinity 1 +554 0 -451.53076 0 -0 0 +555 0 119.92975 1 Infinity 1 +556 0 -136.76703 0 -0 0 +562 0 -604.6978 0 -0 0 +564 0 -561.4162 0 -0 0 +567 0 -411.87683 0 -0 0 +568 1 595.40564 1 -0 1 +570 1 542.1084 1 -0 1 +571 1 1942.8524 1 -0 1 +572 0 -540.9015 0 -0 0 +573 0 -515.32715 0 -0 0 +574 1 1153.9879 1 -0 1 +575 0 -533.9082 0 -0 0 +576 0 -566.47577 0 -0 0 +579 0 -604.6978 0 -0 0 +580 0 -444.53754 0 -0 0 +583 0 -451.53076 0 -0 0 +585 0 -425.95645 0 -0 0 +587 0 -484.83264 0 -0 0 +588 1 962.9796 1 -0 1 +589 0 -477.10504 0 -0 0 +591 1 1292.7548 1 -0 1 +592 1 495.96814 1 -0 1 +595 0 -566.47577 0 -0 0 +596 0 -431.7504 0 -0 0 +597 0 -411.97003 0 -0 0 +598 0 -540.9015 0 -0 0 +599 0 158.95264 1 Infinity 1 +601 0 -385.51306 0 -0 0 +603 1 666.12964 1 -0 1 +605 1 1270.6947 1 -0 1 +608 1 1017.25073 1 -0 1 +610 1 401.65607 1 -0 1 +611 1 1604.3878 1 -0 1 +615 0 -309.81223 0 -0 0 +616 0 -540.9015 0 -0 0 +620 0 -540.9015 0 -0 0 +623 0 -425.95645 0 -0 0 +625 0 -124.750854 0 -0 0 +626 1 592.06555 1 -0 1 +628 0 -438.74362 0 -0 0 +630 0 -105.58676 0 -0 0 +631 0 -566.47577 0 -0 0 +632 0 -425.95645 0 -0 0 +635 0 -85.716156 5.942189E-38 -0 0 +636 1 933.86743 1 -0 1 +637 0 98.515686 1 Infinity 1 +640 0 -389.26315 0 -0 0 +643 0 -425.95645 0 -0 0 +646 0 -163.5892 0 -0 0 +647 0 -350.9017 0 -0 0 +648 1 893.8413 1 -0 1 +650 0 -331.09985 0 -0 0 +651 0 -299.84335 0 -0 0 +655 0 -540.9015 0 -0 0 +658 1 1548.1578 1 -0 1 +659 0 -425.95645 0 -0 0 +662 0 -271.4508 0 -0 0 +663 0 -271.4508 0 -0 0 +664 0 -465.8467 0 -0 0 +666 0 -209.53821 0 -0 0 +667 0 -630.2721 0 -0 0 +669 1 2239.6367 1 -0 1 +671 0 -452.3157 0 -0 0 +672 0 -617.485 0 -0 0 +673 0 -204.12277 0 -0 0 +674 0 -515.32715 0 -0 0 +675 0 -98.50333 1.67E-43 -0 0 +676 0 -393.38895 0 -0 0 +677 0 -477.10504 0 -0 0 +684 0 -425.95645 0 -0 0 +686 0 -425.95645 0 -0 0 +687 0 -377.61526 0 -0 0 +690 0 -350.9017 0 -0 0 +695 0 -438.74362 0 -0 0 diff --git a/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer-out.txt b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer-out.txt index 28ed905c91..7534310404 100644 --- a/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer-out.txt +++ b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer-out.txt @@ -25,23 +25,23 @@ TRUTH ||====================== ||====================== Precision || 0.9585 | 0.9819 | OVERALL 0/1 ACCURACY: 0.973646 -LOG LOSS/instance: 0.112168 +LOG LOSS/instance: 0.110238 Test-set entropy (prior Log-Loss/instance): 0.934003 -LOG-LOSS REDUCTION (RIG): 0.879907 -AUC: 0.996240 +LOG-LOSS REDUCTION (RIG): 0.881972 +AUC: 0.996325 OVERALL RESULTS --------------------------------------- -AUC: 0.996240 (0.0000) +AUC: 0.996325 (0.0000) Accuracy: 0.973646 (0.0000) Positive precision: 0.958506 (0.0000) Positive recall: 0.966527 (0.0000) Negative precision: 0.981900 (0.0000) Negative recall: 0.977477 (0.0000) -Log-loss: 0.112168 (0.0000) -Log-loss reduction: 0.879907 (0.0000) +Log-loss: 0.110238 (0.0000) +Log-loss reduction: 0.881972 (0.0000) F1 Score: 0.962500 (0.0000) -AUPRC: 0.992400 (0.0000) +AUPRC: 0.992721 (0.0000) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer-rp.txt b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer-rp.txt index 5a83bbedba..38d5dd673b 100644 --- a/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer-rp.txt +++ b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer-rp.txt @@ -1,4 +1,4 @@ WeightedEnsemble AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC /bp /nm Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings -0.99624 0.973646 0.958506 0.966527 0.9819 0.977477 0.112168 0.879907 0.9625 0.9924 AvgPer 3 WeightedEnsemble %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsemble{bp=AvgPer nm=3 tp=-} dout=%Output% loader=Text{col=Label:BL:0 col=Features:R4:1-9} data=%Data% out=%Output% seed=1 /bp:AvgPer;/nm:3 +0.996325 0.973646 0.958506 0.966527 0.9819 0.977477 0.110238 0.881972 0.9625 0.992721 AvgPer 3 WeightedEnsemble %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsemble{bp=AvgPer nm=3 tp=-} dout=%Output% loader=Text{col=Label:BL:0 col=Features:R4:1-9} data=%Data% out=%Output% seed=1 /bp:AvgPer;/nm:3 diff --git a/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer.txt b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer.txt index 59ef27a617..f3feea2b57 100644 --- a/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer.txt +++ b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-AvgPer-TrainTest-breast-cancer.txt @@ -1,700 +1,700 @@ Instance Label Score Probability Log-loss Assigned -0 0 -3.482872 0.02253428 0.032881986053906846 0 -1 0 2.9472694 0.89578646 3.2623854205088092 1 -2 0 -4.2678914 0.011065353 0.016052909740335439 0 -3 0 2.986452 0.899107 3.3091017167851975 1 -4 0 -3.5019517 0.022150515 0.032315678223697512 0 -5 1 10.2867985 0.999865 0.00019478350072454816 1 -6 0 -2.3605843 0.060855415 0.090580810904910561 0 -7 0 -4.9923854 0.0057091564 0.0082601734288715676 0 -8 0 -4.683963 0.007570115 0.010962915067788384 0 -9 0 -4.030381 0.013733365 0.019950365620259047 0 -10 0 -5.6011734 0.0032671655 0.0047212402167980012 0 -11 0 -5.476469 0.0036632805 0.0052947005483042842 0 -12 1 -0.42496967 0.27806884 1.8464860048931102 0 -13 0 -5.052911 0.0054013343 0.0078135991288357728 0 -14 1 8.086931 0.9989771 0.0014764523778664522 1 -15 1 1.638257 0.7202814 0.4733674008362686 1 -16 0 -4.338106 0.010379729 0.015053042459133558 0 -17 0 -4.052054 0.013465655 0.019558817434379178 0 -18 1 7.524544 0.99828434 0.0024773000217939069 1 -19 0 -2.9136906 0.037478257 0.055108963848231642 0 -20 1 6.218364 0.99431056 0.0082315687714810456 1 -21 1 7.010026 0.9972474 0.0039766409692087298 1 -22 0 -4.9072876 0.006171662 0.0089314152845014639 0 +0 0 -3.738268 0.019407371 0.028274177879632205 0 +1 0 2.9472685 0.9071517 3.4289806797944715 1 +2 0 -4.4419746 0.010198148 0.014788353232391374 0 +3 0 2.986453 0.91016835 3.4766323525580116 1 +4 0 -3.5572667 0.02287437 0.033384032373821662 0 +5 1 10.030309 0.99985665 0.00020682396203843928 1 +6 0 -2.3605843 0.066329174 0.099014089955006657 0 +7 0 -4.9923854 0.006145395 0.0088932851567994735 0 +8 0 -4.683963 0.008164357 0.011827023914888079 0 +9 0 -4.2586675 0.012065652 0.017512922218736418 0 +10 0 -5.7579794 0.0030301989 0.0043782897664328335 0 +11 0 -5.652498 0.0033406552 0.0048276149428202837 0 +12 1 -0.49986362 0.2852852 1.8095231665202889 0 +13 0 -5.1456814 0.005335127 0.0077175667245616333 0 +14 1 7.5223265 0.9985335 0.0021172834621780222 1 +15 1 1.638257 0.7436518 0.42730081421702953 1 +16 0 -4.556242 0.00918188 0.013307842744969949 0 +17 0 -4.289057 0.011734198 0.017028976282173813 0 +18 1 6.982745 0.9975831 0.0034910827706215904 1 +19 0 -3.187479 0.031936005 0.046825672835095665 0 +20 1 5.8051014 0.9928283 0.010383841783236978 1 +21 1 7.010025 0.99764335 0.0034039375442231515 1 +22 0 -5.10437 0.005542395 0.0080182267005007406 0 23 1 ? ? ? 0 -24 0 -5.7595983 0.002824923 0.0040812697914632779 0 -25 1 1.2282028 0.63836163 0.64755415040416864 1 -26 0 -4.4920645 0.0090201 0.013072299802577814 0 -27 0 -3.7689247 0.017406758 0.025333778099074687 0 -28 0 -5.476469 0.0036632805 0.0052947005483042842 0 -29 0 -5.952343 0.0023665943 0.0034183203055802051 0 -30 0 -4.7488623 0.007134092 0.010329207933888177 0 -31 0 -5.1904163 0.0047619957 0.0068865174780414111 0 -32 1 6.630596 0.99610066 0.0056365490645245509 1 -33 0 -4.8385086 0.0065725357 0.0095134628469589824 0 -34 0 -4.7176075 0.0073408717 0.010629702792356348 0 -35 0 -5.476469 0.0036632805 0.0052947005483042842 0 -36 1 8.456722 0.9992721 0.0010505084408715286 1 -37 0 -1.3033309 0.14642769 0.22841472084732017 0 -38 1 4.9043865 0.9811732 0.027420241139352593 1 -39 1 0.99844265 0.58823794 0.76552825403251523 1 +24 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +25 1 0.897665 0.5933967 0.7529311345918871 1 +26 0 -4.6209464 0.008651597 0.012535923114455177 0 +27 0 -4.0081143 0.015174806 0.022060425445491177 0 +28 0 -5.652498 0.0033406552 0.0048276149428202837 0 +29 0 -6.039849 0.00233462 0.0033720826473683787 0 +30 0 -4.9185214 0.006578372 0.0095219383371659885 0 +31 0 -5.3906345 0.0042553055 0.0061522072207014699 0 +32 1 6.630597 0.9966525 0.0048375462257031088 1 +33 0 -5.0246115 0.0059654796 0.0086321410137232147 0 +34 0 -4.889058 0.0067594172 0.0097848851791912557 0 +35 0 -5.652498 0.0033406552 0.0048276149428202837 0 +36 1 8.000793 0.99905866 0.0013587005999149109 1 +37 0 -1.692574 0.11662218 0.17889748507231804 0 +38 1 4.9043856 0.9836149 0.02383447391076737 1 +39 1 0.99844265 0.6157415 0.69960330804376314 1 40 0 ? ? ? 0 -41 1 3.018794 0.9017767 0.14915790635460019 1 -42 1 7.673623 0.9985041 0.0021597400725719149 1 -43 1 0.637537 0.50608635 0.9825445333891023 1 -44 1 8.207812 0.9990848 0.0013210014038608894 1 -45 0 -5.777409 0.0027790968 0.0040149707997136821 0 -46 1 4.0385695 0.9591516 0.060169195954763985 1 -47 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -48 0 -3.5019517 0.022150515 0.032315678223697512 0 -49 1 5.503627 0.98907024 0.015855120398282622 1 -50 1 3.343049 0.92523605 0.11210662179504392 1 -51 1 -0.16695881 0.32817033 1.6074832876631042 0 -52 1 5.200655 0.98560375 0.02092035007777383 1 -53 1 5.2690516 0.98647064 0.019651982186261987 1 -54 1 5.9823065 0.9929389 0.010223184572271102 1 -55 1 4.7994795 0.97930324 0.030172436055860334 1 -56 1 6.316436 0.9947993 0.007522579396860297 1 -57 1 1.6828299 0.72847575 0.45704714822424003 1 -58 1 1.9818888 0.7794202 0.35952677853476617 1 -59 1 2.06083 0.7916638 0.33704016486982424 1 -60 1 1.6600094 0.72429925 0.46534220942251836 1 -61 0 -5.6923065 0.0030049651 0.0043417750182010864 0 -62 1 7.5410833 0.9983102 0.0024399161320377331 1 -63 1 -0.015551567 0.3596119 1.4754873380652258 0 -64 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -65 1 3.7103176 0.9455205 0.080819326205354647 1 -66 0 -4.052054 0.013465655 0.019558817434379178 0 -67 1 3.3104668 0.9231339 0.1153881546537598 1 -68 1 8.566331 0.99934196 0.00094965672085389622 1 -69 0 -5.5899577 0.0033009714 0.0047701724389436319 0 -70 0 -3.3892312 0.024513878 0.035806746666791144 0 -71 1 6.340682 0.9949136 0.0073568817512015888 1 -72 0 -3.0701742 0.03261282 0.047834674691233356 0 -73 1 7.6628532 0.9984892 0.0021812702713189381 1 -74 1 3.0784302 0.90653473 0.14156580096150426 1 -75 0 -4.5817137 0.008311309 0.012040791132341104 0 -76 0 -5.194137 0.004745785 0.0068630188146811995 0 -77 0 -3.218758 0.028561546 0.041805498203965785 0 -78 0 -3.914548 0.015255628 0.022178828524164255 0 -79 0 -5.6348944 0.0031675892 0.0045771182716585106 0 -80 0 -3.3903775 0.024488648 0.03576943394218645 0 -81 0 -3.986198 0.014295491 0.020772870043326272 0 -82 0 -3.1751475 0.029697128 0.043492951454899474 0 -83 0 -3.1198268 0.031200731 0.045730318686960758 0 -84 1 6.7683554 0.9965636 0.0049661956646408446 1 -85 1 5.2098436 0.9857233 0.020745342488251856 1 -86 1 2.0522451 0.79035693 0.33942375714430012 1 -87 1 4.9122715 0.98130685 0.027223762657824519 1 -88 0 -4.052054 0.013465655 0.019558817434379178 0 -89 0 -5.2829857 0.004374594 0.0063250498864140005 0 -90 0 -5.7595983 0.002824923 0.0040812697914632779 0 -91 0 -4.8414316 0.0065549845 0.0094879744610039916 0 -92 0 -4.052054 0.013465655 0.019558817434379178 0 -93 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -94 0 -5.1904163 0.0047619957 0.0068865174780414111 0 -95 0 -5.7595983 0.002824923 0.0040812697914632779 0 -96 0 -5.410613 0.0038914173 0.0056250803074518237 0 -97 0 -3.482872 0.02253428 0.032881986053906846 0 -98 1 8.737397 0.9994378 0.00081129825513571107 1 -99 1 9.01284 0.9995637 0.00062959385993691976 1 -100 1 4.706623 0.9774973 0.032835407454128211 1 -101 1 -1.1443181 0.16569202 2.5934240038475713 0 -102 0 -3.8623738 0.01599446 0.023261655961929854 0 -103 1 1.3917551 0.6723575 0.57269956165021552 1 -104 1 10.23489 0.9998584 0.00020432985822878773 1 -105 1 2.2230587 0.815232 0.2947174500496586 1 -106 1 9.76918 0.9997825 0.00031381647486206597 1 -107 1 5.137453 0.9847542 0.022164422947398725 1 -108 0 -5.7063694 0.0029664172 0.0042859956037871088 0 -109 1 5.4784746 0.98881704 0.016224495496090417 1 -110 0 -3.3888607 0.024522038 0.035818815315306611 0 -111 1 3.3876276 0.92802656 0.10776200391078569 1 -112 1 8.628671 0.9993786 0.0008967382033695237 1 -113 1 8.882472 0.9995081 0.00070986101950228368 1 -114 0 -3.1094184 0.031491753 0.046163760805275879 0 -115 0 -4.9183283 0.006109615 0.008841347254921127 0 -116 0 -0.25753403 0.3100516 0.53543960504643839 0 -117 1 7.6684456 0.99849695 0.002170074527871256 1 -118 0 -5.8763533 0.0025376931 0.0036657705638862108 0 -119 0 -3.960167 0.014637227 0.021273127467339784 0 -120 0 -5.21801 0.0046430747 0.0067141400647920081 0 -121 0 -3.390985 0.024475288 0.035749674704110175 0 -122 1 9.622348 0.99975103 0.00035923048673869361 1 -123 1 3.4594145 0.9323189 0.10110454124506661 1 -124 1 6.0973387 0.9936441 0.0091988635063169621 1 -125 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -126 1 7.136181 0.9975485 0.003541079440976193 1 -127 0 -4.621235 0.008016654 0.011612195307125778 0 -128 1 4.493952 0.9727618 0.039841504527236671 1 -129 0 -6.1793385 0.0019210464 0.0027741496163762847 0 -130 0 -3.3892312 0.024513878 0.035806746666791144 0 -131 0 -5.1904163 0.0047619957 0.0068865174780414111 0 -132 1 7.395813 0.9980688 0.0027888126758143891 1 -133 0 -4.737647 0.007207617 0.010436048402163285 0 -134 0 -5.065713 0.005338371 0.0077222716426920144 0 -135 0 -2.7902474 0.04180148 0.061603507815141713 0 -136 0 -4.338106 0.010379729 0.015053042459133558 0 -137 0 -5.19334 0.0047492543 0.0068680476530536229 0 -138 0 -4.1484265 0.0123362485 0.01790813268936867 0 +41 1 2.7321234 0.8889206 0.16987352580346732 1 +42 1 7.673623 0.9987253 0.0018401835805509819 1 +43 1 0.637537 0.5341261 0.90474770393703829 1 +44 1 8.207812 0.999223 0.0011214186493592648 1 +45 0 -5.920701 0.0026067493 0.0037656544842424168 0 +46 1 4.0385695 0.9641426 0.052681522331917578 1 +47 0 -6.200626 0.0020117955 0.0029053308785867587 0 +48 0 -3.5572667 0.02287437 0.033384032373821662 0 +49 1 5.503627 0.99053603 0.013718635893576811 1 +50 1 3.343049 0.93379426 0.098823374436285291 1 +51 1 -0.19723034 0.3457766 1.5320879064073591 0 +52 1 5.200657 0.9875031 0.018142801150648202 1 +53 1 5.2034616 0.9875352 0.018095953114010289 1 +54 1 5.9823065 0.9939088 0.0088145844557812984 1 +55 1 4.7994814 0.9819705 0.026248426739313425 1 +56 1 6.316436 0.99552494 0.0064706317681734772 1 +57 1 1.6828299 0.7514544 0.41224250934101159 1 +58 1 1.9818888 0.7996017 0.32264660231428366 1 +59 1 1.665534 0.7484457 0.418030462738193 1 +60 1 1.5219374 0.72254723 0.46883619483305711 1 +61 0 -5.856407 0.0027665084 0.0039967590012408329 0 +62 1 7.5410833 0.99855876 0.0020807700542075469 1 +63 1 -0.015551567 0.38482422 1.3777285064561582 0 +64 0 -6.200626 0.0020117955 0.0029053308785867587 0 +65 1 3.2980146 0.93116426 0.10289240193295103 1 +66 0 -4.289057 0.011734198 0.017028976282173813 0 +67 1 3.3104677 0.93190104 0.10175133744346095 1 +68 1 8.566331 0.9994427 0.00080424301730056371 1 +69 0 -5.7938833 0.0029312272 0.0042350769814508164 0 +70 0 -3.3892312 0.026630158 0.038940018135035295 0 +71 1 5.7077494 0.9921558 0.011361421565394849 1 +72 0 -3.0701742 0.035477098 0.05211260042044711 0 +73 1 7.184512 0.99799484 0.0028957383591989841 1 +74 1 2.5785675 0.87405705 0.19420063943377938 1 +75 0 -4.5817137 0.00896939 0.012998475968654409 0 +76 0 -5.1941376 0.005101817 0.0073792060928969932 0 +77 0 -3.2187586 0.031051008 0.045507374520449714 0 +78 0 -4.0441036 0.014683868 0.0213414168390797 0 +79 0 -5.7559705 0.003035834 0.0043864443746869401 0 +80 0 -3.5975199 0.022054462 0.032173971123134658 0 +81 0 -4.185153 0.012906194 0.018740900372745485 0 +82 0 -3.4537416 0.025122134 0.036706607486316573 0 +83 0 -3.1198268 0.03393427 0.049806741833892498 0 +84 1 6.7683554 0.9970529 0.0042580325880637974 1 +85 1 5.2098436 0.98760784 0.01798981049554501 1 +86 1 2.052246 0.8098559 0.30426290551373475 1 +87 1 4.860862 0.98295134 0.02480809120460643 1 +88 0 -4.289057 0.011734198 0.017028976282173813 0 +89 0 -5.4672575 0.0039644977 0.0057309288308588667 0 +90 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +91 0 -4.990079 0.006158476 0.0089122743860173392 0 +92 0 -4.289057 0.011734198 0.017028976282173813 0 +93 0 -6.200626 0.0020117955 0.0029053308785867587 0 +94 0 -5.3906345 0.0042553055 0.0061522072207014699 0 +95 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +96 0 -5.538207 0.0037129077 0.0053665624295177157 0 +97 0 -3.738268 0.019407371 0.028274177879632205 0 +98 1 8.737397 0.9995245 0.00068620196061278886 1 +99 1 8.389928 0.9993437 0.00094716133251951037 1 +100 1 4.706624 0.9803804 0.028586429677332305 1 +101 1 -1.4226594 0.14499266 2.7859481823721555 0 +102 0 -3.890808 0.01688985 0.024575025807886251 0 +103 1 1.3917551 0.69769984 0.51932158208889256 1 +104 1 9.46399 0.9997576 0.00034976911637384886 1 +105 1 1.9302044 0.79180855 0.33677645715690274 1 +106 1 9.2541895 0.9997055 0.00042494571667690465 1 +107 1 5.137452 0.9867585 0.019231096328357004 1 +108 0 -5.9007363 0.0026553473 0.0038359515931879642 0 +109 1 4.849698 0.98277694 0.02506408855791066 1 +110 0 -3.5399618 0.023235932 0.033917965672728707 0 +111 1 3.3794346 0.9358507 0.095649736648938707 1 +112 1 8.272346 0.9992681 0.0010562740679413579 1 +113 1 8.882472 0.9995843 0.00059982818132431012 1 +114 0 -3.2419453 0.030410353 0.044553800818252581 0 +115 0 -5.0805492 0.0056655295 0.0081968731426666202 0 +116 0 -0.37149525 0.31017247 0.53569238581060752 0 +117 1 7.269723 0.99814695 0.0026758643363441331 1 +118 0 -5.8763533 0.002715928 0.0039235863818907089 0 +119 0 -4.2057786 0.0126647 0.01838798629009674 0 +120 0 -5.319969 0.0045422893 0.0065680665978747783 0 +121 0 -3.657651 0.020882918 0.030446707842719321 0 +122 1 9.622346 0.99979067 0.00030203314854602157 1 +123 1 3.4594135 0.9401633 0.089016707962115263 1 +124 1 6.0973387 0.9945219 0.0079249313395690565 1 +125 0 -6.200626 0.0020117955 0.0029053308785867587 0 +126 1 6.597887 0.9965497 0.0049863008723421021 1 +127 0 -4.8398457 0.007072917 0.010240319857589488 0 +128 1 4.493951 0.9762033 0.034746433078818091 1 +129 0 -6.1793385 0.0020518356 0.0029632141673322301 0 +130 0 -3.3892312 0.026630158 0.038940018135035295 0 +131 0 -5.3906345 0.0042553055 0.0061522072207014699 0 +132 1 6.8712063 0.99732035 0.003871100926986012 1 +133 0 -4.9544253 0.0063642506 0.0092110145070213294 0 +134 0 -5.2051816 0.0050500785 0.0073041821173898281 0 +135 0 -3.1072083 0.034320112 0.050383064200888034 0 +136 0 -4.556242 0.00918188 0.013307842744969949 0 +137 0 -5.361168 0.004372712 0.0063223231877638552 0 +138 0 -4.3382697 0.011216421 0.016273310211744527 0 139 0 ? ? ? 0 -140 0 -5.19334 0.0047492543 0.0068680476530536229 0 -141 0 -5.7625213 0.002817351 0.0040703146280809182 0 -142 1 4.097539 0.9612269 0.057051101127774083 1 -143 0 -4.9183283 0.006109615 0.008841347254921127 0 -144 0 -5.476469 0.0036632805 0.0052947005483042842 0 +140 0 -5.361168 0.004372712 0.0063223231877638552 0 +141 0 -5.9092956 0.0026344028 0.0038056549166740083 0 +142 1 3.724495 0.9525894 0.070073610891559848 1 +143 0 -5.0805492 0.0056655295 0.0081968731426666202 0 +144 0 -5.652498 0.0033406552 0.0048276149428202837 0 145 0 ? ? ? 0 -146 1 1.1432829 0.620121 0.68937834352130667 1 -147 0 -5.850732 0.0025981204 0.0037531730783225315 0 -148 0 -2.192862 0.07030446 0.10516976150026086 0 -149 1 9.191527 0.99962986 0.00053410498513138422 1 -150 0 -5.6011734 0.0032671655 0.0047212402167980012 0 -151 1 4.351611 0.96906507 0.045334552844361685 1 -152 1 9.273413 0.99965674 0.0004953090586974708 1 -153 0 -3.9174714 0.015215238 0.022119655786370637 0 -154 0 -6.331703 0.0016699884 0.002411297942606464 0 -155 1 3.2879562 0.9216501 0.11770893420530119 1 -156 0 -5.5482464 0.0034297842 0.0049566376322899286 0 -157 0 -5.1904163 0.0047619957 0.0068865174780414111 0 +146 1 1.1432838 0.646999 0.62816460914221506 1 +147 0 -5.9088883 0.0026353956 0.0038070909936062211 0 +148 0 -2.192862 0.076639555 0.11503416409636115 0 +149 1 9.191529 0.9996879 0.00045032085391415828 1 +150 0 -5.7579794 0.0030301989 0.0043782897664328335 0 +151 1 4.351611 0.9729352 0.039584374290501351 1 +152 1 8.432736 0.9993692 0.00091033334453149658 1 +153 0 -4.1246023 0.013641708 0.019816297395817094 0 +154 0 -6.457423 0.0015860328 0.0022899781237268467 0 +155 1 2.9329042 0.90602326 0.14238000010439222 1 +156 0 -5.7214613 0.0031342814 0.0045289133535339841 0 +157 0 -5.3906345 0.0042553055 0.0061522072207014699 0 158 0 ? ? ? 0 -159 1 10.745463 0.9999115 0.00012770277002010949 1 -160 1 8.114793 0.999003 0.0014390944132832524 1 -161 0 -3.8824134 0.015706606 0.022839683107138374 0 -162 0 -4.621235 0.008016654 0.011612195307125778 0 -163 0 -3.5052323 0.022085175 0.032219280817519881 0 +159 1 10.745463 0.99992615 0.00010654718652104918 1 +160 1 8.114794 0.999153 0.0012224545059740068 1 +161 0 -4.1494994 0.013334405 0.019366891376986155 0 +162 0 -4.8398457 0.007072917 0.010240319857589488 0 +163 0 -3.5325022 0.023393506 0.034150723870132307 0 164 0 ? ? ? 0 -165 0 -3.5290694 0.021616044 0.031527348116069484 0 -166 1 6.9524107 0.99709785 0.0041930049474674319 1 -167 1 8.171879 0.999054 0.0013654142587147774 1 -168 0 -4.621235 0.008016654 0.011612195307125778 0 -169 0 -5.8213696 0.002669137 0.0038558989253168036 0 -170 0 -5.19334 0.0047492543 0.0068680476530536229 0 -171 0 -5.7595983 0.002824923 0.0040812697914632779 0 -172 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -173 1 12.098251 0.99997455 3.6718763229178855E-05 1 -174 1 5.4575663 0.9886021 0.016538121186036488 1 -175 1 6.7321196 0.99644744 0.0051343804750916197 1 -176 0 -5.1904163 0.0047619957 0.0068865174780414111 0 -177 1 5.343604 0.9873569 0.018356423098867477 1 -178 0 -4.052054 0.013465655 0.019558817434379178 0 -179 1 1.817894 0.752367 0.41049148582615391 1 -180 0 -5.6011734 0.0032671655 0.0047212402167980012 0 -181 0 -6.331703 0.0016699884 0.002411297942606464 0 -182 0 -2.9136906 0.037478257 0.055108963848231642 0 -183 1 7.498831 0.9982433 0.0025366510899867789 1 -184 1 5.5104933 0.9891384 0.015755662663344983 1 -185 0 -5.031992 0.005505811 0.0079651536563163441 0 -186 1 4.5500593 0.97409785 0.037861396865713125 1 -187 1 11.832129 0.99996746 4.6952027696279339E-05 1 -188 1 7.640403 0.9984577 0.0022268292311602884 1 -189 0 -4.94297 0.005973361 0.0086435796330161952 0 -190 1 10.813717 0.99991685 0.00011996288641172571 1 -191 1 10.09983 0.9998396 0.0002314212167623026 1 -192 0 -3.7689247 0.017406758 0.025333778099074687 0 -193 0 -5.7595983 0.002824923 0.0040812697914632779 0 -194 0 -4.621235 0.008016654 0.011612195307125778 0 -195 0 -4.052054 0.013465655 0.019558817434379178 0 -196 0 5.527667 0.989307 6.5471876771068054 1 -197 0 -2.9226785 0.037180834 0.054663233597824101 0 -198 0 -6.331703 0.0016699884 0.002411297942606464 0 -199 0 -4.9072876 0.006171662 0.0089314152845014639 0 -200 1 8.758909 0.99944884 0.00079538099915120008 1 -201 1 8.602963 0.9993638 0.00091816351562154204 1 -202 0 -5.7595983 0.002824923 0.0040812697914632779 0 -203 0 -3.482872 0.02253428 0.032881986053906846 0 -204 0 -5.7595983 0.002824923 0.0040812697914632779 0 -205 1 10.099371 0.99983954 0.00023150722188518337 1 -206 1 6.829487 0.9967511 0.0046947596442636786 1 -207 0 -5.6011734 0.0032671655 0.0047212402167980012 0 -208 0 -5.6011734 0.0032671655 0.0047212402167980012 0 -209 0 -3.3692598 0.024957517 0.036463016082005542 0 -210 1 12.871542 0.9999875 1.805829135842506E-05 1 -211 1 9.103922 0.9995988 0.00057892375102128462 1 -212 0 -5.7595983 0.002824923 0.0040812697914632779 0 -213 1 12.87052 0.9999875 1.805829135842506E-05 1 -214 1 12.879089 0.9999876 1.7886306565083914E-05 1 -215 1 6.851302 0.99681556 0.004601503047228275 1 -216 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -217 0 -5.7595983 0.002824923 0.0040812697914632779 0 -218 1 6.756961 0.9965275 0.0050184870456513927 1 -219 0 -2.3909802 0.05927525 0.088155434815330072 0 -220 0 -5.40769 0.0039018649 0.0056402119369919996 0 -221 1 8.626977 0.99937767 0.00089811492069533539 1 -222 1 -2.4763036 0.0550428 4.1833023401596607 0 -223 1 4.5353794 0.9737546 0.038369878966271906 1 -224 1 8.90344 0.9995175 0.00069626776733784027 1 -225 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -226 1 8.736739 0.99943745 0.00081181449340512405 1 -227 1 7.0285454 0.9972938 0.0039095566169136797 1 -228 0 -5.6011734 0.0032671655 0.0047212402167980012 0 -229 1 11.740766 0.9999646 5.1079751554952729E-05 1 -230 1 5.1917076 0.9854864 0.021092150358962274 1 -231 1 6.7423306 0.9964806 0.0050863996385837502 1 -232 0 1.7194624 0.73509645 1.9164609425071408 1 -233 1 5.4949408 0.98898345 0.015981712887257544 1 -234 0 -3.126328 0.031020278 0.045461620615876844 0 +165 0 -3.7750583 0.01876842 0.027334429135731558 0 +166 1 6.9524117 0.9975143 0.0035905606102989286 1 +167 1 7.822379 0.9988894 0.0016031667808534577 1 +168 0 -4.8398457 0.007072917 0.010240319857589488 0 +169 0 -5.900486 0.0026559625 0.0038368414135946404 0 +170 0 -5.361168 0.004372712 0.0063223231877638552 0 +171 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +172 0 -6.200626 0.0020117955 0.0029053308785867587 0 +173 1 11.258599 0.9999541 6.621484007793289E-05 1 +174 1 4.903079 0.9835954 0.02386314919255644 1 +175 1 6.7321186 0.9969525 0.0044033635734041501 1 +176 0 -5.3906345 0.0042553055 0.0061522072207014699 0 +177 1 5.343603 0.9890381 0.015901982668991399 1 +178 0 -4.289057 0.011734198 0.017028976282173813 0 +179 1 1.817894 0.7741119 0.3693859184239649 1 +180 0 -5.7579794 0.0030301989 0.0043782897664328335 0 +181 0 -6.457423 0.0015860328 0.0022899781237268467 0 +182 0 -3.187479 0.031936005 0.046825672835095665 0 +183 1 7.024873 0.99767554 0.0033573932890881958 1 +184 1 5.0216036 0.9852783 0.021396799329704182 1 +185 0 -5.2098513 0.0050283596 0.0072726896589810673 0 +186 1 4.5500593 0.97738296 0.033004145527381921 1 +187 1 11.22812 0.9999528 6.8106737307562526E-05 1 +188 1 6.9585485 0.9975284 0.0035702162307141108 1 +189 0 -4.9429703 0.006431802 0.0093090984382543963 0 +190 1 10.454881 0.9999033 0.00013948467255915295 1 +191 1 10.099832 0.9998656 0.00019392347161872577 1 +192 0 -4.0081143 0.015174806 0.022060425445491177 0 +193 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +194 0 -4.8398457 0.007072917 0.010240319857589488 0 +195 0 -4.289057 0.011734198 0.017028976282173813 0 +196 0 5.00879 0.9851049 6.069020170363574 1 +197 0 -2.9226785 0.040468488 0.059597907934809839 0 +198 0 -6.457423 0.0015860328 0.0022899781237268467 0 +199 0 -5.10437 0.005542395 0.0080182267005007406 0 +200 1 8.407546 0.9993543 0.00093184490556906757 1 +201 1 8.602963 0.9994613 0.00077739901311388369 1 +202 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +203 0 -3.738268 0.019407371 0.028274177879632205 0 +204 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +205 1 9.380837 0.99973816 0.00037780935818556969 1 +206 1 6.829487 0.9972149 0.004023636364508496 1 +207 0 -5.7579794 0.0030301989 0.0043782897664328335 0 +208 0 -5.7579794 0.0030301989 0.0043782897664328335 0 +209 0 -3.5654678 0.022704942 0.033133898887283193 0 +210 1 12.11888 0.9999793 2.9839298506057709E-05 1 +211 1 8.404681 0.99935263 0.00093425422037926294 1 +212 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +213 1 12.0267935 0.9999775 3.2419093932940516E-05 1 +214 1 12.1010275 0.999979 3.0269264090181833E-05 1 +215 1 6.851302 0.9972706 0.0039430984031390282 1 +216 0 -6.200626 0.0020117955 0.0029053308785867587 0 +217 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +218 1 5.9937325 0.99397266 0.008721926306471145 1 +219 0 -2.3909798 0.064604186 0.096351121598510864 0 +220 0 -5.5727396 0.0035962737 0.0051976780049017672 0 +221 1 8.626977 0.99947315 0.00076027761748214545 1 +222 1 -2.476303 0.05998277 4.0593080619281201 0 +223 1 4.5353785 0.9770799 0.033451511302562423 1 +224 1 8.180217 0.99920285 0.0011505066118902478 1 +225 0 -6.200626 0.0020117955 0.0029053308785867587 0 +226 1 8.736739 0.9995242 0.0006866321218568387 1 +227 1 6.456297 0.99606735 0.0056848071940564888 1 +228 0 -5.7579794 0.0030301989 0.0043782897664328335 0 +229 1 11.740766 0.9999707 4.2308352471765759E-05 1 +230 1 5.1917086 0.9874003 0.018293021192954294 1 +231 1 6.7423296 0.9969811 0.0043619621575598806 1 +232 0 1.7194624 0.75774693 2.0454131754185476 1 +233 1 5.4949408 0.99046016 0.013829152973616962 1 +234 0 -3.3749237 0.02697636 0.039453238611965251 0 235 0 ? ? ? 0 -236 1 9.951018 0.99981606 0.00026539363927387412 1 -237 1 6.384386 0.9951132 0.0070674535345721205 1 -238 1 11.298662 0.99994683 7.6706301438007131E-05 1 -239 1 5.218775 0.9858386 0.020576636425471754 1 -240 0 -1.993123 0.08331826 0.12550716430778644 0 -241 0 -4.1647773 0.012154139 0.017642146609719373 0 -242 0 -5.1904163 0.0047619957 0.0068865174780414111 0 -243 0 -3.4550157 0.023106266 0.03372645951020347 0 -244 0 -5.7595983 0.002824923 0.0040812697914632779 0 -245 0 -2.7776203 0.042269703 0.062308653918616327 0 -246 1 10.201185 0.9998539 0.00021078013554602304 1 -247 1 3.0554304 0.90472466 0.14444930340846809 1 -248 0 -3.2038512 0.028944885 0.042374913183321754 0 +236 1 9.951016 0.9998457 0.0002226487211611543 1 +237 1 6.384387 0.9957972 0.0060761115996594483 1 +238 1 10.682345 0.9999217 0.0001129970266666251 1 +239 1 4.6218424 0.97880906 0.030900641577567965 1 +240 0 -2.2586493 0.072430916 0.1084733595212949 0 +241 0 -4.1647773 0.013149223 0.019096145947227944 0 +242 0 -5.3906345 0.0042553055 0.0061522072207014699 0 +243 0 -3.4550161 0.025093192 0.036663777974401762 0 +244 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +245 0 -3.0360093 0.0365777 0.053759779365932077 0 +246 1 9.549155 0.999776 0.00032319159885338563 1 +247 1 3.0554304 0.91526455 0.1277392965218232 1 +248 0 -3.5497203 0.023031361 0.033615843506398714 0 249 0 ? ? ? 0 -250 0 -5.834298 0.0026376322 0.0038103262134336524 0 -251 1 8.333951 0.9991851 0.0011761526984772684 1 -252 0 3.7254658 0.94623464 4.2171793115221385 1 -253 1 7.673623 0.9985041 0.0021597400725719149 1 -254 1 7.5410833 0.9983102 0.0024399161320377331 1 -255 1 3.8116865 0.9501337 0.073797583784302623 1 -256 0 -5.19334 0.0047492543 0.0068680476530536229 0 -257 0 -4.9072876 0.006171662 0.0089314152845014639 0 -258 0 -4.621235 0.008016654 0.011612195307125778 0 -259 0 3.242813 0.9185952 3.6187422330022576 1 -260 1 9.219126 0.99963915 0.00052068543357584364 1 -261 1 9.322227 0.9996718 0.0004735459469822395 1 -262 1 8.625552 0.99937683 0.00089931954943311293 1 -263 1 7.8194847 0.99869186 0.0018884870944413687 1 -264 1 4.7406597 0.9781765 0.031833327029850804 1 -265 0 -2.4273925 0.05743295 0.085332845290028464 0 -266 1 7.700018 0.9985399 0.0021079827894124729 1 -267 1 1.9541712 0.77500075 0.36773038644864792 1 -268 1 8.424179 0.99925 0.0010824348145564958 1 -269 0 -5.7595983 0.002824923 0.0040812697914632779 0 -270 1 5.901844 0.9924 0.011006373382357104 1 -271 0 -3.482872 0.02253428 0.032881986053906846 0 -272 1 1.9541712 0.77500075 0.36773038644864792 1 -273 1 0.26800632 0.42166668 1.2458250641387991 1 -274 0 -4.451595 0.009359383 0.013566321171082689 0 +250 0 -5.978259 0.0024715506 0.0035701074650224761 0 +251 1 8.237916 0.9992444 0.0010905240888066275 1 +252 0 3.311242 0.93194664 3.877189695297437 1 +253 1 7.673623 0.9987253 0.0018401835805509819 1 +254 1 7.5410833 0.99855876 0.0020807700542075469 1 +255 1 3.8116865 0.95611143 0.064749326306391991 1 +256 0 -5.361168 0.004372712 0.0063223231877638552 0 +257 0 -5.10437 0.005542395 0.0080182267005007406 0 +258 0 -4.8398457 0.007072917 0.010240319857589488 0 +259 0 3.242814 0.9278087 3.7920312457007515 1 +260 1 8.776113 0.9995412 0.0006620271048482008 1 +261 1 9.322227 0.9997235 0.00039896891985811181 1 +262 1 8.625552 0.9994725 0.00076122402098052809 1 +263 1 7.234562 0.9980856 0.0027645164055080333 1 +264 1 4.185087 0.9685563 0.046092205487729791 1 +265 0 -2.4273925 0.06259272 0.093252099720831205 0 +266 1 7.066353 0.9977631 0.0032307832747264201 1 +267 1 1.9541712 0.79544985 0.33015711240175655 1 +268 1 7.7044973 0.99876124 0.0017882655641669536 1 +269 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +270 1 5.901844 0.99343985 0.009495471276816889 1 +271 0 -3.738268 0.019407371 0.028274177879632205 0 +272 1 1.9541712 0.79544985 0.33015711240175655 1 +273 1 0.26800632 0.44866192 1.1562993428781159 1 +274 0 -4.6976275 0.008062348 0.011678651336700034 0 275 0 ? ? ? 0 -276 0 -4.9072876 0.006171662 0.0089314152845014639 0 -277 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -278 0 -5.7595983 0.002824923 0.0040812697914632779 0 -279 1 6.614441 0.9960425 0.0057208076013373407 1 -280 0 -4.621235 0.008016654 0.011612195307125778 0 -281 0 -4.8385086 0.0065725357 0.0095134628469589824 0 -282 1 2.6904383 0.8715527 0.19834018320850152 1 -283 1 5.7269773 0.991084 0.012920785377970624 1 -284 1 6.3651476 0.9950263 0.0071934500825941066 1 -285 1 11.654726 0.9999617 5.5293481842253808E-05 1 -286 1 13.146536 0.9999903 1.401665413523095E-05 1 -287 0 -5.065713 0.005338371 0.0077222716426920144 0 -288 1 1.3048286 0.6544869 0.61156379083602264 1 -289 1 6.631445 0.9961037 0.0056321463459408586 1 -290 0 -6.331703 0.0016699884 0.002411297942606464 0 -291 0 -5.7595983 0.002824923 0.0040812697914632779 0 +276 0 -5.10437 0.005542395 0.0080182267005007406 0 +277 0 -6.200626 0.0020117955 0.0029053308785867587 0 +278 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +279 1 6.150161 0.9947825 0.007546955930147493 1 +280 0 -4.8398457 0.007072917 0.010240319857589488 0 +281 0 -5.0246115 0.0059654796 0.0086321410137232147 0 +282 1 2.6414633 0.8803409 0.18386584126392186 1 +283 1 5.7269754 0.9922934 0.011161311660882557 1 +284 1 6.3651495 0.9957218 0.0061853538663619624 1 +285 1 10.793361 0.99992937 0.0001019033194709417 1 +286 1 12.527284 0.9999859 2.0380088075209094E-05 1 +287 0 -5.2051816 0.0050500785 0.0073041821173898281 0 +288 1 1.3048277 0.680426 0.55548982221541199 1 +289 1 6.3977356 0.9958487 0.0060015034537152416 1 +290 0 -6.457423 0.0015860328 0.0022899781237268467 0 +291 0 -5.9414234 0.0025572449 0.0036940497744623112 0 292 1 ? ? ? 0 -293 1 5.297948 0.9868212 0.019139422425450404 1 +293 1 5.297947 0.98856926 0.016586049475147857 1 294 0 ? ? ? 0 -295 1 5.721692 0.9910409 0.012983517781355288 1 -296 0 1.2067633 0.6337916 1.4492633116388667 1 +295 1 5.721692 0.9922558 0.011215994634413492 1 +296 0 1.2067633 0.660329 1.5577899771500425 1 297 0 ? ? ? 0 -298 0 -2.5149827 0.053219307 0.078897808656333432 0 -299 1 6.0366135 0.993281 0.0097261698992318259 1 -300 1 7.189887 0.99766654 0.0033704082906736892 1 -301 0 -5.7595983 0.002824923 0.0040812697914632779 0 -302 1 13.107393 0.9999899 1.4532607192850369E-05 1 -303 0 -5.7595983 0.002824923 0.0040812697914632779 0 -304 1 5.193618 0.9855115 0.021055415313673605 1 -305 1 8.285159 0.99914765 0.0012302003065924249 1 -306 0 -5.7595983 0.002824923 0.0040812697914632779 0 -307 0 -5.7595983 0.002824923 0.0040812697914632779 0 -308 1 6.519494 0.9956824 0.0062424394797278623 1 -309 0 -2.4410834 0.056754258 0.08429441173243564 0 -310 0 -5.6348944 0.0031675892 0.0045771182716585106 0 -311 0 -6.331703 0.0016699884 0.002411297942606464 0 -312 1 3.018118 0.90172154 0.14924611489916423 1 -313 0 -6.331703 0.0016699884 0.002411297942606464 0 -314 0 -5.8872256 0.0025124764 0.0036292984276296014 0 +298 0 -2.5324926 0.057110284 0.084839056539765381 0 +299 1 5.4112816 0.9896983 0.014939307746220086 1 +300 1 7.189887 0.9980048 0.0028813490265551982 1 +301 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +302 1 12.611642 0.99998695 1.8832223182178005E-05 1 +303 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +304 1 5.193618 0.9874223 0.018260885851386126 1 +305 1 7.6459284 0.99869215 0.0018880565746969143 1 +306 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +307 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +308 1 5.86306 0.99320114 0.0098421823955161681 1 +309 0 -2.4410834 0.06185165 0.092112020548319032 0 +310 0 -5.7559705 0.003035834 0.0043864443746869401 0 +311 0 -6.457423 0.0015860328 0.0022899781237268467 0 +312 1 3.018118 0.9125413 0.13203828773629589 1 +313 0 -6.457423 0.0015860328 0.0022899781237268467 0 +314 0 -6.014777 0.002389423 0.0034513338255624817 0 315 0 ? ? ? 0 -316 1 3.2541513 0.9193725 0.12127858314360404 1 -317 1 7.468912 0.99819434 0.0026073759429630396 1 -318 0 -5.695045 0.0029974198 0.0043308566215409967 0 -319 0 1.4018602 0.6744041 1.6188454941410018 1 -320 1 5.6245403 0.99021065 0.014192625478375142 1 +316 1 3.2541513 0.92850995 0.1070107235969851 1 +317 1 6.899477 0.99738955 0.0037710002272127274 1 +318 0 -5.6950455 0.0032117849 0.0046410831880226919 0 +319 0 1.4018602 0.6996733 1.7353953301896854 1 +320 1 4.98415 0.98476577 0.022147482456960178 1 321 0 ? ? ? 0 -322 0 -4.621235 0.008016654 0.011612195307125778 0 -323 1 4.8959904 0.98102987 0.027631033936048165 1 -324 0 -5.7595983 0.002824923 0.0040812697914632779 0 -325 0 -4.070602 0.013240634 0.019229787788375932 0 -326 1 3.624156 0.9412861 0.087294823984529965 1 -327 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -328 1 3.361329 0.92639214 0.11030508520262793 1 -329 1 6.433284 0.99532735 0.0067570028317258763 1 -330 1 5.3615503 0.9875616 0.01805737864221952 1 -331 0 -3.2146301 0.028667204 0.041962421527380285 0 -332 0 -2.749971 0.043312535 0.063880398900094393 0 -333 1 3.9098234 0.9542453 0.067567875521126988 1 -334 1 5.6489344 0.990426 0.013878901445224296 1 -335 0 -6.331703 0.0016699884 0.002411297942606464 0 -336 1 4.8692474 0.9805661 0.028313232035807549 1 -337 0 -5.7595983 0.002824923 0.0040812697914632779 0 -338 0 -5.8872256 0.0025124764 0.0036292984276296014 0 -339 1 5.427353 0.9882843 0.017001987752258311 1 -340 1 5.5303516 0.9893331 0.015471758985912978 1 -341 0 -5.7595983 0.002824923 0.0040812697914632779 0 -342 0 -5.7625213 0.002817351 0.0040703146280809182 0 -343 0 -6.331703 0.0016699884 0.002411297942606464 0 -344 1 7.070565 0.99739623 0.0037613440240078367 1 -345 0 -6.331703 0.0016699884 0.002411297942606464 0 -346 0 -3.024808 0.033956848 0.049840461040150705 0 -347 0 -5.833997 0.0026383623 0.0038113823933699754 0 -348 1 -0.55175304 0.2552492 1.9700216427351105 0 -349 1 2.3500576 0.83220255 0.26499337935318013 1 -350 0 -3.3673954 0.024999332 0.036524887125210473 0 -351 0 -5.1904163 0.0047619957 0.0068865174780414111 0 -352 0 0.03989792 0.37145302 0.66990750678422029 1 -353 1 7.090433 0.9974433 0.0036932351406289404 1 -354 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -355 0 -4.326891 0.010486351 0.015208487929545276 0 -356 1 -0.6269388 0.24231228 2.045060560925192 0 -357 1 10.086641 0.99983764 0.00023425938852449468 1 -358 1 5.204255 0.9856507 0.020851600795979936 1 -359 1 4.969612 0.9822512 0.025836030173023942 1 -360 1 12.014021 0.99997246 3.9728539361084329E-05 1 -361 1 5.618044 0.9901525 0.014277385220010059 1 -362 0 -3.727109 0.01807774 0.026319286432117856 0 -363 0 -1.7872968 0.09898463 0.15037637664232212 0 -364 0 -5.1904163 0.0047619957 0.0068865174780414111 0 -365 0 -5.476469 0.0036632805 0.0052947005483042842 0 -366 1 11.240093 0.99994385 8.1006102725794237E-05 1 -367 1 9.364462 0.99968433 0.0004554819533959377 1 -368 0 -5.952343 0.0023665943 0.0034183203055802051 0 -369 0 -5.507866 0.0035592567 0.0051440819427372234 0 -370 0 -3.691402 0.018670749 0.02719083081176573 0 -371 0 -5.952343 0.0023665943 0.0034183203055802051 0 -372 0 -4.1484265 0.0123362485 0.01790813268936867 0 -373 0 -3.4170165 0.023909425 0.034913068495921094 0 -374 0 -4.7176075 0.0073408717 0.010629702792356348 0 -375 0 -6.331703 0.0016699884 0.002411297942606464 0 -376 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -377 0 -5.8872256 0.0025124764 0.0036292984276296014 0 -378 0 -3.8403473 0.01631684 0.023734390824775076 0 -379 0 -1.9753585 0.08457621 0.12748831270224811 0 -380 0 -6.331703 0.0016699884 0.002411297942606464 0 -381 1 9.040222 0.99957454 0.00061393669231691098 1 -382 0 -3.6022816 0.020235304 0.02949278685454829 0 -383 0 -5.7625213 0.002817351 0.0040703146280809182 0 -384 0 -5.7625213 0.002817351 0.0040703146280809182 0 -385 0 -3.488603 0.022418328 0.032710857018100691 0 -386 1 5.1706343 0.9852062 0.021502407370847236 1 -387 0 -2.2325287 0.067954145 0.10152716080481058 0 -388 0 -5.3068285 0.0042800005 0.0061879871982490689 0 -389 0 -3.4043655 0.024182815 0.035317204424318911 0 -390 0 -5.6937428 0.0030010056 0.0043360454354717873 0 -391 1 9.233915 0.99964404 0.00051363161678326592 1 -392 0 -4.9072876 0.006171662 0.0089314152845014639 0 -393 0 -6.173278 0.0019317768 0.0027896601179374578 0 -394 0 -5.3585663 0.0040816837 0.005900675385287716 0 -395 0 -4.9072876 0.006171662 0.0089314152845014639 0 -396 0 -4.621235 0.008016654 0.011612195307125778 0 -397 0 -4.6241584 0.007995274 0.011581101010143969 0 -398 0 -4.4299226 0.009546243 0.013838475255114937 0 -399 0 -5.6446185 0.0031394402 0.0045363793898473706 0 -400 1 7.713833 0.9985584 0.0020812867469309085 1 -401 0 -5.19334 0.0047492543 0.0068680476530536229 0 -402 0 -2.774115 0.042400565 0.062505793817949698 0 -403 0 -4.133309 0.0125070205 0.01815760338927197 0 -404 0 -5.1505213 0.0049393116 0.0071435771586442244 0 -405 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -406 0 -3.8407016 0.016311606 0.023726714477872276 0 -407 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -408 0 -3.680035 0.018863497 0.027474226469704165 0 -409 0 -4.7176075 0.0073408717 0.010629702792356348 0 -410 0 -6.0456505 0.0021721658 0.0031371812770679881 0 +322 0 -4.8398457 0.007072917 0.010240319857589488 0 +323 1 4.7439194 0.98103493 0.02762358335373118 1 +324 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +325 0 -4.325316 0.011350477 0.016468919916732855 0 +326 1 3.3639822 0.9349847 0.096985362520202636 1 +327 0 -6.200626 0.0020117955 0.0029053308785867587 0 +328 1 3.3186588 0.9323817 0.10100742231351764 1 +329 1 6.433284 0.9959828 0.0058072295118437008 1 +330 1 5.3615503 0.9892171 0.015640912257682788 1 +331 0 -3.3378692 0.027893443 0.04081363155082831 0 +332 0 -3.0079165 0.037507236 0.055152400307940143 0 +333 1 3.9098234 0.9597768 0.059229126462331604 1 +334 1 5.6489344 0.9917196 0.01199582081404962 1 +335 0 -6.457423 0.0015860328 0.0022899781237268467 0 +336 1 4.487195 0.9760573 0.03496226363571129 1 +337 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +338 0 -6.014777 0.002389423 0.0034513338255624817 0 +339 1 4.9487343 0.98426497 0.022881345182366477 1 +340 1 5.219531 0.9877174 0.017829784132127061 1 +341 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +342 0 -5.9092956 0.0026344028 0.0038056549166740083 0 +343 0 -6.457423 0.0015860328 0.0022899781237268467 0 +344 1 6.738328 0.9969699 0.004378177570557586 1 +345 0 -6.457423 0.0015860328 0.0022899781237268467 0 +346 0 -3.3006182 0.028845884 0.042227834380300894 0 +347 0 -5.9716854 0.0024866306 0.0035919173390103684 0 +348 1 -0.55175304 0.2755728 1.859494564332927 0 +349 1 2.1060982 0.8174297 0.29083339487040744 1 +350 0 -3.5227346 0.023601409 0.034457881730402082 0 +351 0 -5.3906345 0.0042553055 0.0061522072207014699 0 +352 0 0.03989792 0.39707115 0.72994033825056359 1 +353 1 7.090433 0.9978124 0.0031595107755524859 1 +354 0 -6.200626 0.0020117955 0.0029053308785867587 0 +355 0 -4.538265 0.009334842 0.013530581523950659 0 +356 1 -0.6269388 0.26186925 1.9330814242164462 0 +357 1 10.028682 0.9998564 0.00020716797669539043 1 +358 1 5.204255 0.98754424 0.018082717513100623 1 +359 1 4.9306307 0.9840027 0.02326580694855097 1 +360 1 11.28114 0.99995506 6.4838916378310478E-05 1 +361 1 5.241521 0.98796237 0.017472009195776982 1 +362 0 -3.7271094 0.019605344 0.028565474836170109 0 +363 0 -1.9169598 0.096829005 0.14692893898450018 0 +364 0 -5.3906345 0.0042553055 0.0061522072207014699 0 +365 0 -5.652498 0.0033406552 0.0048276149428202837 0 +366 1 10.488206 0.99990624 0.00013527069636485589 1 +367 1 8.676394 0.99949676 0.00072620750501173098 1 +368 0 -6.039849 0.00233462 0.0033720826473683787 0 +369 0 -5.651598 0.0033434355 0.0048316394178647276 0 +370 0 -3.8522344 0.017494397 0.025462460489090878 0 +371 0 -6.039849 0.00233462 0.0033720826473683787 0 +372 0 -4.3382697 0.011216421 0.016273310211744527 0 +373 0 -3.6370254 0.02127774 0.031028583201373111 0 +374 0 -4.889058 0.0067594172 0.0097848851791912557 0 +375 0 -6.457423 0.0015860328 0.0022899781237268467 0 +376 0 -6.200626 0.0020117955 0.0029053308785867587 0 +377 0 -6.014777 0.002389423 0.0034513338255624817 0 +378 0 -4.0270076 0.014915097 0.021680020599020907 0 +379 0 -1.975359 0.092193685 0.1395435714644632 0 +380 0 -6.457423 0.0015860328 0.0022899781237268467 0 +381 1 8.338668 0.99931175 0.00099328369012960664 1 +382 0 -3.847374 0.017572066 0.02557651225669156 0 +383 0 -5.9092956 0.0026344028 0.0038056549166740083 0 +384 0 -5.9092956 0.0026344028 0.0038056549166740083 0 +385 0 -3.6893945 0.02028924 0.029572210240656815 0 +386 1 4.608096 0.9785429 0.031292958620012731 1 +387 0 -2.2325287 0.07407582 0.11103402971718652 0 +388 0 -5.502553 0.0038372849 0.0055466807612930477 0 +389 0 -3.5896053 0.02221337 0.032408415864608818 0 +390 0 -5.829537 0.0028361357 0.0040974921556309477 0 +391 1 8.47907 0.9993957 0.00087204355947112706 1 +392 0 -5.10437 0.005542395 0.0080182267005007406 0 +393 0 -6.271575 0.0018838919 0.00272044478546917 0 +394 0 -5.3904324 0.0042561004 0.0061533588968344854 0 +395 0 -5.10437 0.005542395 0.0080182267005007406 0 +396 0 -4.8398457 0.007072917 0.010240319857589488 0 +397 0 -4.81304 0.0072497097 0.010497217242063378 0 +398 0 -4.6568503 0.008370562 0.012126994935525031 0 +399 0 -5.796669 0.0029236844 0.0042241630980287886 0 +400 1 7.1387615 0.9979081 0.0030211125696380688 1 +401 0 -5.361168 0.004372712 0.0063223231877638552 0 +402 0 -2.985322 0.03827127 0.056298078305914719 0 +403 0 -4.328512 0.011317253 0.016420438330730858 0 +404 0 -5.267726 0.004766771 0.006893439841980405 0 +405 0 -6.200626 0.0020117955 0.0029053308785867587 0 +406 0 -4.077078 0.014247811 0.02070308643002346 0 +407 0 -6.200626 0.0020117955 0.0029053308785867587 0 +408 0 -3.9036531 0.01669313 0.024286372581299141 0 +409 0 -4.889058 0.0067594172 0.0097848851791912557 0 +410 0 -6.200626 0.0020117955 0.0029053308785867587 0 411 0 ? ? ? 0 -412 1 8.314893 0.99917066 0.001196979721687337 1 -413 0 -3.2931924 0.026720002 0.039073188964596009 0 -414 1 5.8874807 0.99229956 0.011152385793831057 1 -415 0 -1.409133 0.13466503 0.20866938414777755 0 -416 1 7.729929 0.99857956 0.0020507160792223671 1 -417 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -418 0 -1.8940396 0.09055709 0.13694502039538939 0 -419 0 -5.622347 0.0032042828 0.0046302252048672264 0 -420 0 -2.7726307 0.0424561 0.062589466320259954 0 -421 1 10.155396 0.9998476 0.00021989657667089834 1 -422 0 -2.7179637 0.044550397 0.065748316609816643 0 -423 0 -3.3892312 0.024513878 0.035806746666791144 0 -424 0 -5.19334 0.0047492543 0.0068680476530536229 0 -425 1 12.06453 0.9999737 3.7922672928455309E-05 1 -426 0 -2.2380357 0.067633666 0.10103118264460387 0 -427 1 4.6389713 0.9760854 0.034920680707081007 1 -428 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -429 0 -5.476469 0.0036632805 0.0052947005483042842 0 -430 0 -5.260561 0.00446546 0.0064567236482734533 0 -431 0 -2.9262657 0.03706276 0.054486322913896434 0 -432 0 -3.5992842 0.0202901 0.029573474708830459 0 -433 0 -3.9921188 0.014218866 0.020660723932839499 0 -434 0 4.5476446 0.9740417 5.2676603051217787 1 -435 1 7.575901 0.99836344 0.0023629979497649372 1 -436 1 5.1114473 0.98439044 0.022697451404313559 1 -437 0 -4.6241584 0.007995274 0.011581101010143969 0 -438 0 -3.5409827 0.021385249 0.031187065341240661 0 -439 0 -4.054977 0.013429944 0.019506595612928591 0 -440 1 8.698355 0.99941725 0.00084098225569382591 1 -441 0 -1.7145677 0.10512017 0.16023412890252972 0 -442 0 -4.9211993 0.006093582 0.0088180749336334366 0 -443 0 -5.978359 0.002310701 0.0033374942601122489 0 -444 0 -2.7438335 0.04354732 0.064234502814603439 0 -445 0 -5.7625213 0.002817351 0.0040703146280809182 0 -446 0 -6.331703 0.0016699884 0.002411297942606464 0 -447 0 -4.054977 0.013429944 0.019506595612928591 0 -448 0 -6.173278 0.0019317768 0.0027896601179374578 0 -449 1 8.28597 0.99914825 0.0012293396600255831 1 -450 0 -3.8532417 0.01612735 0.02345650577373825 0 -451 0 -4.054977 0.013429944 0.019506595612928591 0 -452 0 -4.4895234 0.00904104 0.013102784928163535 0 -453 1 5.9675274 0.99284285 0.010362708492085076 1 -454 0 -5.305497 0.0042852284 0.006195562018052281 0 -455 1 1.2593336 0.64495295 0.63273417024068512 1 -456 1 9.107021 0.99959993 0.00057728926100807164 1 -457 1 7.5455313 0.9983171 0.0024299242886964066 1 -458 0 -3.8652973 0.015952144 0.023199616520614492 0 -459 0 -3.6756177 0.01893893 0.027585150464716924 0 -460 0 -3.3673954 0.024999332 0.036524887125210473 0 -461 0 -3.7799654 0.017233718 0.025079734435049596 0 -462 0 -2.8788795 0.03865192 0.056869206933533242 0 -463 0 -4.454518 0.00933446 0.013530025451277884 0 -464 0 -4.6241584 0.007995274 0.011581101010143969 0 -465 1 8.897833 0.999515 0.00069988115098498695 1 -466 1 7.822398 0.9986954 0.0018834069696430824 1 -467 1 6.104863 0.9936877 0.0091356031502358923 1 -468 0 -4.6241584 0.007995274 0.011581101010143969 0 -469 0 -5.8559704 0.00258565 0.0037351353289297126 0 -470 0 -4.7488623 0.007134092 0.010329207933888177 0 -471 0 -2.8788795 0.03865192 0.056869206933533242 0 -472 0 -3.3273702 0.025913604 0.037878357359513184 0 -473 0 -4.6241584 0.007995274 0.011581101010143969 0 -474 0 -4.054977 0.013429944 0.019506595612928591 0 -475 0 -5.19334 0.0047492543 0.0068680476530536229 0 -476 0 -4.4344788 0.009506655 0.013780812917959804 0 -477 0 -4.6241584 0.007995274 0.011581101010143969 0 -478 0 -3.7472525 0.017751403 0.025839893107689336 0 -479 1 7.0932884 0.99745 0.0036835794579032264 1 -480 0 -3.896552 0.015506589 0.022546544157841086 0 -481 0 -2.716064 0.04462492 0.065860850250515399 0 -482 1 12.526953 0.99998283 2.4765714290683422E-05 1 -483 1 8.983756 0.99955183 0.00064671370472097975 1 -484 0 -3.8652973 0.015952144 0.023199616520614492 0 -485 0 -4.8073 0.0067628636 0.0097898910700309032 0 -486 0 -4.7488623 0.007134092 0.010329207933888177 0 -487 1 12.056477 0.99997354 3.8180653709006879E-05 1 -488 1 0.5997734 0.497393 1.0075418558412388 1 -489 1 -0.7618656 0.22023608 2.1828772719853711 0 -490 0 -6.331703 0.0016699884 0.002411297942606464 0 -491 1 5.410244 0.98810047 0.017270353764690164 1 -492 0 -4.1796813 0.011990457 0.017403118056768217 0 -493 1 8.512299 0.9993084 0.00099810252898520769 1 -494 0 -0.02289772 0.35805553 0.63947959376297381 0 -495 0 -4.7488623 0.007134092 0.010329207933888177 0 -496 0 -6.173278 0.0019317768 0.0027896601179374578 0 -497 0 -4.3164334 0.010586746 0.01535486923496476 0 -498 0 -4.338106 0.010379729 0.015053042459133558 0 -499 0 -4.338106 0.010379729 0.015053042459133558 0 -500 0 -2.9136906 0.037478257 0.055108963848231642 0 -501 0 -4.338106 0.010379729 0.015053042459133558 0 -502 0 -3.986198 0.014295491 0.020772870043326272 0 -503 0 -4.052054 0.013465655 0.019558817434379178 0 -504 0 -6.331703 0.0016699884 0.002411297942606464 0 -505 0 -4.5778904 0.008340377 0.012083080118329342 0 -506 1 9.572725 0.9997394 0.00037600306851729912 1 -507 0 -5.271672 0.004420206 0.0063911447218310408 0 -508 0 -4.054977 0.013429944 0.019506595612928591 0 -509 0 -5.7625213 0.002817351 0.0040703146280809182 0 -510 0 -6.331703 0.0016699884 0.002411297942606464 0 -511 0 -3.7689247 0.017406758 0.025333778099074687 0 -512 0 -4.054977 0.013429944 0.019506595612928591 0 -513 0 -4.7488623 0.007134092 0.010329207933888177 0 -514 1 9.177126 0.9996249 0.00054124492561566752 1 -515 1 7.4553976 0.99817175 0.0026400259790533786 1 -516 0 -6.173278 0.0019317768 0.0027896601179374578 0 -517 0 -5.8872256 0.0025124764 0.0036292984276296014 0 -518 0 -4.6517515 0.007796244 0.011291676640520974 0 -519 1 4.6810236 0.9769728 0.033609671221477525 1 -520 0 -6.490128 0.0014436262 0.0020842171987779552 0 -521 0 -4.782583 0.0069174776 0.010014488179386138 0 -522 1 4.839095 0.9800299 0.029102357379861676 1 -523 1 6.776593 0.99658954 0.0049286609408553195 1 -524 0 -4.9072876 0.006171662 0.0089314152845014639 0 -525 0 -4.8414316 0.0065549845 0.0094879744610039916 0 -526 0 -4.6241584 0.007995274 0.011581101010143969 0 -527 0 -4.052054 0.013465655 0.019558817434379178 0 -528 0 -2.754528 0.04313898 0.063618701593183336 0 -529 0 -4.1796813 0.011990457 0.017403118056768217 0 -530 1 5.300787 0.98685515 0.019089753637405036 1 -531 0 -3.8407016 0.016311606 0.023726714477872276 0 -532 0 -5.6011734 0.0032671655 0.0047212402167980012 0 -533 0 -4.9072876 0.006171662 0.0089314152845014639 0 -534 0 -5.476469 0.0036632805 0.0052947005483042842 0 -535 0 -4.921719 0.0060906843 0.0088138686520135151 0 -536 0 -3.482872 0.02253428 0.032881986053906846 0 -537 0 -3.2931924 0.026720002 0.039073188964596009 0 -538 0 -4.338106 0.010379729 0.015053042459133558 0 -539 0 -3.1997428 0.029051412 0.042533188174337766 0 -540 0 -3.4155807 0.0239403 0.034958703821406929 0 -541 0 -5.19334 0.0047492543 0.0068680476530536229 0 -542 0 -3.4864955 0.022460902 0.032773688933512055 0 -543 0 -4.338106 0.010379729 0.015053042459133558 0 -544 0 -4.745201 0.0071580117 0.010363965031894386 0 -545 0 -3.7689247 0.017406758 0.025333778099074687 0 -546 1 10.264345 0.99986213 0.00019891164756910478 1 -547 0 -5.6040964 0.0032584118 0.0047085699080465756 0 -548 0 -5.034915 0.0054910923 0.0079438017710837285 0 -549 1 5.6715956 0.99062186 0.013593630704674639 1 -550 0 -4.9072876 0.006171662 0.0089314152845014639 0 -551 0 -5.7595983 0.002824923 0.0040812697914632779 0 -552 0 -3.3162293 0.026173826 0.038263818197394765 0 -553 0 -1.7405233 0.102892905 0.15664787393522983 0 -554 0 -5.19334 0.0047492543 0.0068680476530536229 0 -555 0 -1.963068 0.08545659 0.12887643973743862 0 -556 0 -2.926115 0.037067715 0.054493746084402689 0 -557 0 -3.3673954 0.024999332 0.036524887125210473 0 -558 0 -5.476469 0.0036632805 0.0052947005483042842 0 -559 0 -3.7689247 0.017406758 0.025333778099074687 0 -560 0 -3.482872 0.02253428 0.032881986053906846 0 -561 0 -3.482872 0.02253428 0.032881986053906846 0 -562 0 -5.7595983 0.002824923 0.0040812697914632779 0 -563 0 -4.9072876 0.006171662 0.0089314152845014639 0 -564 0 -3.8824134 0.015706606 0.022839683107138374 0 -565 1 9.47172 0.999714 0.00041264538703320383 1 -566 0 -4.4315553 0.009532038 0.013817785080896181 0 -567 0 -3.6314187 0.019710151 0.028719711784198444 0 -568 1 4.041501 0.95925725 0.060010338652604994 1 -569 1 9.748814 0.9997784 0.00031975117930321329 1 -570 1 6.448406 0.99539167 0.0066637856180221805 1 -571 1 9.226857 0.9996417 0.00051698647656674178 1 -572 0 -4.9072876 0.006171662 0.0089314152845014639 0 -573 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -574 1 5.8652763 0.9921417 0.011381876111811766 1 -575 0 -3.2931924 0.026720002 0.039073188964596009 0 -576 0 -3.7689247 0.017406758 0.025333778099074687 0 -577 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -578 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -579 0 -5.7595983 0.002824923 0.0040812697914632779 0 -580 0 -3.5792446 0.020660188 0.030118560948251851 0 -581 1 7.14419 0.9975665 0.0035150464757207813 1 -582 1 7.248555 0.99778897 0.0031933798554606417 1 -583 0 -5.19334 0.0047492543 0.0068680476530536229 0 -584 0 -2.9208136 0.03724236 0.054755428877105654 0 -585 0 -6.331703 0.0016699884 0.002411297942606464 0 -586 1 11.9440155 0.9999706 4.2394346321571357E-05 1 -587 0 -3.5992842 0.0202901 0.029573474708830459 0 -588 1 5.449994 0.98852324 0.016653203948794274 1 -589 0 -4.054977 0.013429944 0.019506595612928591 0 -590 1 3.138379 0.9111083 0.13430551922296985 1 -591 1 4.1688643 0.96360195 0.053490786532823395 1 -592 1 5.2628794 0.9863945 0.019763303451007887 1 -593 0 -3.8652973 0.015952144 0.023199616520614492 0 -594 1 3.7616892 0.94790655 0.07718325210685581 1 -595 0 -3.7689247 0.017406758 0.025333778099074687 0 -596 0 -4.1484265 0.0123362485 0.01790813268936867 0 -597 0 -2.9417691 0.036556605 0.053728188627819098 0 -598 0 -4.9072876 0.006171662 0.0089314152845014639 0 -599 0 -3.0197873 0.034108836 0.050067458929324682 0 -600 0 -4.9072876 0.006171662 0.0089314152845014639 0 -601 0 -5.8872256 0.0025124764 0.0036292984276296014 0 -602 0 -4.338106 0.010379729 0.015053042459133558 0 -603 1 3.4922361 0.9342013 0.098194641924081588 1 -604 1 4.188938 0.9642448 0.052528659757042101 1 -605 1 9.458142 0.99971044 0.0004178063517362282 1 -606 0 -4.27225 0.011021516 0.015988960936891675 0 -607 0 -6.331703 0.0016699884 0.002411297942606464 0 -608 1 8.117621 0.9990056 0.0014353070238773355 1 -609 0 -4.054977 0.013429944 0.019506595612928591 0 -610 1 8.266865 0.99913317 0.0012511141759918355 1 -611 1 6.1478367 0.9939311 0.0087822269655119054 1 -612 1 14.358191 0.99999684 4.5575474466196295E-06 1 -613 0 -4.861019 0.0064385654 0.0093189189633460356 0 -614 0 -5.3180437 0.0042362125 0.0061245443306950425 0 -615 0 -3.94002 0.014907228 0.0216684965636894 0 -616 0 -4.9072876 0.006171662 0.0089314152845014639 0 +412 1 8.314893 0.9992964 0.0010153988510166489 1 +413 0 -3.3400192 0.027839413 0.040733448691679393 0 +414 1 5.8874807 0.9933524 0.0096224591623897213 1 +415 0 -1.409133 0.14655516 0.22863017556539333 0 +416 1 7.729929 0.9987901 0.0017465947434439401 1 +417 0 -6.200626 0.0020117955 0.0029053308785867587 0 +418 0 -2.0004616 0.09026317 0.13647883888523057 0 +419 0 -5.635271 0.0033942864 0.0049052497580990865 0 +420 0 -2.9377928 0.039927542 0.058784803063903983 0 +421 1 10.155396 0.9998723 0.00018420517835190389 1 +422 0 -2.8563938 0.042924788 0.06329579093541482 0 +423 0 -3.3892312 0.026630158 0.038940018135035295 0 +424 0 -5.361168 0.004372712 0.0063223231877638552 0 +425 1 12.06453 0.99997824 3.1387175208611342E-05 1 +426 0 -2.2380362 0.07372616 0.11048932919007244 0 +427 1 4.6389723 0.97913617 0.030418584784864745 1 +428 0 -6.200626 0.0020117955 0.0029053308785867587 0 +429 0 -5.652498 0.0033406552 0.0048276149428202837 0 +430 0 -5.455531 0.0040076855 0.0057934849468162103 0 +431 0 -3.011478 0.037388146 0.054973905418033188 0 +432 0 -3.8581696 0.017400013 0.025323875300941854 0 +433 0 -4.139826 0.013452985 0.019540289530273286 0 +434 0 4.2261915 0.969697 5.0443954952203169 1 +435 1 7.575902 0.9986045 0.0020147210248404315 1 +436 1 5.1114464 0.9864395 0.019697486003574433 1 +437 0 -4.81304 0.0072497097 0.010497217242063378 0 +438 0 -3.7076826 0.019954747 0.029079727940932894 0 +439 0 -4.2649117 0.0119968 0.01741238050521976 0 +440 1 8.108878 0.99914837 0.0012291675307738255 1 +441 0 -1.7145672 0.11453672 0.17549561095093449 0 +442 0 -5.0268087 0.0059534046 0.0086146159407858251 0 +443 0 -6.113205 0.0021813714 0.0031504912068339688 0 +444 0 -2.743834 0.04742514 0.070095620290486493 0 +445 0 -5.9092956 0.0026344028 0.0038056549166740083 0 +446 0 -6.457423 0.0015860328 0.0022899781237268467 0 +447 0 -4.2649117 0.0119968 0.01741238050521976 0 +448 0 -6.271575 0.0018838919 0.00272044478546917 0 +449 1 8.28597 0.9992773 0.001043021765479907 1 +450 0 -3.8532414 0.017478347 0.025438892519199494 0 +451 0 -4.2649117 0.0119968 0.01741238050521976 0 +452 0 -4.61899 0.0086671775 0.012558596780817194 0 +453 1 5.4320784 0.98989314 0.014655304047687313 1 +454 0 -5.4999943 0.0038463688 0.0055598365832204245 0 +455 1 1.1107101 0.6400674 0.64370426722541996 1 +456 1 8.37531 0.99933475 0.00096006856013463606 1 +457 1 7.5455294 0.99856466 0.0020722446509886341 1 +458 0 -4.0833225 0.01416668 0.020584351909383104 0 +459 0 -3.8366933 0.017743928 0.025828914411809498 0 +460 0 -3.5227346 0.023601409 0.034457881730402082 0 +461 0 -3.9842935 0.015508605 0.02254949890416191 0 +462 0 -2.9746065 0.03863882 0.056849545127545635 0 +463 0 -4.6630945 0.008322619 0.01205724483116118 0 +464 0 -4.81304 0.0072497097 0.010497217242063378 0 +465 1 8.189908 0.99921 0.0011401794574540424 1 +466 1 7.822397 0.99888945 0.0016030806939212961 1 +467 1 5.6803427 0.99195546 0.011652752856256334 1 +468 0 -4.81304 0.0072497097 0.010497217242063378 0 +469 0 -5.990636 0.002443405 0.0035294020597964296 0 +470 0 -4.9185214 0.006578372 0.0095219383371659885 0 +471 0 -2.9746065 0.03863882 0.056849545127545635 0 +472 0 -3.5309353 0.023426736 0.034199813221724495 0 +473 0 -4.81304 0.0072497097 0.010497217242063378 0 +474 0 -4.2649117 0.0119968 0.01741238050521976 0 +475 0 -5.361168 0.004372712 0.0063223231877638552 0 +476 0 -4.6314497 0.008568429 0.012414894945345718 0 +477 0 -4.81304 0.0072497097 0.010497217242063378 0 +478 0 -3.9673367 0.015750611 0.022904183390089507 0 +479 1 6.5661564 0.996447 0.0051349845605430794 1 +480 0 -4.0790634 0.014221965 0.020665259986327789 0 +481 0 -2.716064 0.048602574 0.071879972137842529 0 +482 1 11.649966 0.9999681 4.600609264199643E-05 1 +483 1 8.983756 0.99962157 0.00054606225482140512 1 +484 0 -4.0833225 0.01416668 0.020584351909383104 0 +485 0 -4.84096 0.007065661 0.010229776551850263 0 +486 0 -4.9185214 0.006578372 0.0095219383371659885 0 +487 1 11.399861 0.99995977 5.8045312346478289E-05 1 +488 1 0.59977245 0.52539927 0.92851390531143441 1 +489 1 -0.7618656 0.23840562 2.0685098793515171 0 +490 0 -6.457423 0.0015860328 0.0022899781237268467 0 +491 1 5.2613935 0.9881797 0.017154699644096223 1 +492 0 -4.3703933 0.010890694 0.015798134037447245 0 +493 1 7.905631 0.9989719 0.0014840273826560388 1 +494 0 -0.022896767 0.38321242 0.69715437413825143 0 +495 0 -4.9185214 0.006578372 0.0095219383371659885 0 +496 0 -6.271575 0.0018838919 0.00272044478546917 0 +497 0 -4.515465 0.009532482 0.013818432153053737 0 +498 0 -4.556242 0.00918188 0.013307842744969949 0 +499 0 -4.556242 0.00918188 0.013307842744969949 0 +500 0 -3.187479 0.031936005 0.046825672835095665 0 +501 0 -4.556242 0.00918188 0.013307842744969949 0 +502 0 -4.185153 0.012906194 0.018740900372745485 0 +503 0 -4.289057 0.011734198 0.017028976282173813 0 +504 0 -6.457423 0.0015860328 0.0022899781237268467 0 +505 0 -4.766018 0.0075704926 0.010963464059485761 0 +506 1 8.978934 0.9996199 0.00054847092545639903 1 +507 0 -5.4247675 0.0041232253 0.0059608541890844042 0 +508 0 -4.2649117 0.0119968 0.01741238050521976 0 +509 0 -5.9092956 0.0026344028 0.0038056549166740083 0 +510 0 -6.457423 0.0015860328 0.0022899781237268467 0 +511 0 -4.0081143 0.015174806 0.022060425445491177 0 +512 0 -4.2649117 0.0119968 0.01741238050521976 0 +513 0 -4.9185214 0.006578372 0.0095219383371659885 0 +514 1 8.539557 0.9994287 0.00082446238870957883 1 +515 1 6.9524384 0.99751437 0.0035904744046954444 1 +516 0 -6.271575 0.0018838919 0.00272044478546917 0 +517 0 -6.014777 0.002389423 0.0034513338255624817 0 +518 0 -4.808489 0.007280157 0.010541464813451511 0 +519 1 4.6810236 0.97991836 0.029266534950097269 1 +520 0 -6.6432724 0.0013352035 0.0019275786545103757 0 +521 0 -4.998888 0.006108662 0.0088399642885958838 0 +522 1 4.3081856 0.971854 0.041188540775102388 1 +523 1 6.3227396 0.99555093 0.0064329715078171798 1 +524 0 -5.10437 0.005542395 0.0080182267005007406 0 +525 0 -4.990079 0.006158476 0.0089122743860173392 0 +526 0 -4.81304 0.0072497097 0.010497217242063378 0 +527 0 -4.289057 0.011734198 0.017028976282173813 0 +528 0 -2.9825153 0.038367216 0.056442012933679335 0 +529 0 -4.3703933 0.010890694 0.015798134037447245 0 +530 1 5.300788 0.988599 0.016542644295914406 1 +531 0 -4.077078 0.014247811 0.02070308643002346 0 +532 0 -5.7579794 0.0030301989 0.0043782897664328335 0 +533 0 -5.10437 0.005542395 0.0080182267005007406 0 +534 0 -5.652498 0.0033406552 0.0048276149428202837 0 +535 0 -4.9217196 0.006559012 0.0094938232723815966 0 +536 0 -3.738268 0.019407371 0.028274177879632205 0 +537 0 -3.3400192 0.027839413 0.040733448691679393 0 +538 0 -4.556242 0.00918188 0.013307842744969949 0 +539 0 -3.4599862 0.02498065 0.036497243402783416 0 +540 0 -3.6638956 0.0207648 0.030272675891029326 0 +541 0 -5.361168 0.004372712 0.0063223231877638552 0 +542 0 -3.6697617 0.02065444 0.030110093239691042 0 +543 0 -4.556242 0.00918188 0.013307842744969949 0 +544 0 -4.8183928 0.007214058 0.010445408332707573 0 +545 0 -4.0081143 0.015174806 0.022060425445491177 0 +546 1 9.552054 0.9997766 0.00032233149319667721 1 +547 0 -5.723447 0.0031285316 0.004520592156712566 0 +548 0 -5.1753187 0.0051911967 0.0075088205059808504 0 +549 1 5.3795795 0.98939407 0.015382844124239577 1 +550 0 -5.10437 0.005542395 0.0080182267005007406 0 +551 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +552 0 -3.452847 0.025142465 0.036736694752149572 0 +553 0 -1.7766018 0.10882866 0.16622525255432827 0 +554 0 -5.361168 0.004372712 0.0063223231877638552 0 +555 0 -2.397884 0.06421823 0.095755971952985658 0 +556 0 -2.9720125 0.038728304 0.056983839982703767 0 +557 0 -3.5227346 0.023601409 0.034457881730402082 0 +558 0 -5.652498 0.0033406552 0.0048276149428202837 0 +559 0 -4.0081143 0.015174806 0.022060425445491177 0 +560 0 -3.738268 0.019407371 0.028274177879632205 0 +561 0 -3.738268 0.019407371 0.028274177879632205 0 +562 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +563 0 -5.10437 0.005542395 0.0080182267005007406 0 +564 0 -4.1494994 0.013334405 0.019366891376986155 0 +565 1 9.145295 0.9996742 0.00047010516886122363 1 +566 0 -4.441596 0.0102016935 0.014793521084320157 0 +567 0 -3.8678045 0.017247858 0.025100491050177667 0 +568 1 4.041501 0.9642365 0.052541055826792971 1 +569 1 9.009376 0.99963045 0.0005332447537232187 1 +570 1 5.646085 0.99169785 0.012027470060416649 1 +571 1 9.226857 0.9996979 0.00043586987358929348 1 +572 0 -5.10437 0.005542395 0.0080182267005007406 0 +573 0 -6.200626 0.0020117955 0.0029053308785867587 0 +574 1 5.3970995 0.9895633 0.015136118874829655 1 +575 0 -3.3400192 0.027839413 0.040733448691679393 0 +576 0 -4.0081143 0.015174806 0.022060425445491177 0 +577 0 -6.200626 0.0020117955 0.0029053308785867587 0 +578 0 -6.200626 0.0020117955 0.0029053308785867587 0 +579 0 -5.9414234 0.0025572449 0.0036940497744623112 0 +580 0 -3.7874808 0.018557368 0.027024153967500603 0 +581 1 7.144189 0.9979186 0.0030059464501394459 1 +582 1 6.531864 0.9963326 0.0053006861202848079 1 +583 0 -5.361168 0.004372712 0.0063223231877638552 0 +584 0 -2.920814 0.0405357 0.059698967007212007 0 +585 0 -6.457423 0.0015860328 0.0022899781237268467 0 +586 1 11.047006 0.99994415 8.0576122020335265E-05 1 +587 0 -3.8581696 0.017400013 0.025323875300941854 0 +588 1 4.9658966 0.98450965 0.022522752187804532 1 +589 0 -4.2649117 0.0119968 0.01741238050521976 0 +590 1 2.5600986 0.87215894 0.19733701527529834 1 +591 1 4.1688643 0.9680947 0.046779904861229085 1 +592 1 4.7825127 0.98168963 0.026661116413936772 1 +593 0 -4.0833225 0.01416668 0.020584351909383104 0 +594 1 3.3619509 0.93487006 0.097162233284273009 1 +595 0 -4.0081143 0.015174806 0.022060425445491177 0 +596 0 -4.3382697 0.011216421 0.016273310211744527 0 +597 0 -2.9417696 0.03978637 0.058572677947965689 0 +598 0 -5.10437 0.005542395 0.0080182267005007406 0 +599 0 -3.2367196 0.030553613 0.04476697906499099 0 +600 0 -5.10437 0.005542395 0.0080182267005007406 0 +601 0 -6.014777 0.002389423 0.0034513338255624817 0 +602 0 -4.556242 0.00918188 0.013307842744969949 0 +603 1 3.2511377 0.92832416 0.10729942462881266 1 +604 1 3.727378 0.95271003 0.069890913683330405 1 +605 1 9.458144 0.9997563 0.0003516613854829185 1 +606 0 -4.441951 0.010198371 0.014788677664894058 0 +607 0 -6.457423 0.0015860328 0.0022899781237268467 0 +608 1 8.117621 0.9991552 0.0012192701333380953 1 +609 0 -4.2649117 0.0119968 0.01741238050521976 0 +610 1 7.8042593 0.9988706 0.0016302844201503833 1 +611 1 5.5714035 0.9911076 0.01288642687749666 1 +612 1 13.566325 0.9999946 7.8252318363944361E-06 1 +613 0 -5.0573483 0.005788074 0.0083746855736656304 0 +614 0 -5.466649 0.0039667273 0.0057341582505026972 0 +615 0 -3.94002 0.016148308 0.023487238431928348 0 +616 0 -5.10437 0.005542395 0.0080182267005007406 0 617 0 ? ? ? 0 -618 0 -4.338106 0.010379729 0.015053042459133558 0 -619 0 -3.7689247 0.017406758 0.025333778099074687 0 -620 0 -4.9072876 0.006171662 0.0089314152845014639 0 -621 0 -0.16198826 0.32918024 0.57600290997485726 0 -622 0 -1.7370653 0.10318721 0.15712124281408327 0 -623 0 -6.331703 0.0016699884 0.002411297942606464 0 -624 0 -3.4554825 0.023096565 0.033712133470186707 0 -625 0 -3.6625314 0.019164136 0.02791636266616257 0 -626 1 4.185855 0.96414673 0.052675368274755011 1 -627 0 -4.349147 0.010275814 0.014901560170357533 0 -628 0 -5.7625213 0.002817351 0.0040703146280809182 0 -629 0 -4.6241584 0.007995274 0.011581101010143969 0 -630 0 -2.7987113 0.04149041 0.061135227872361926 0 -631 0 -3.7689247 0.017406758 0.025333778099074687 0 -632 0 -6.331703 0.0016699884 0.002411297942606464 0 -633 1 3.9723816 0.9566957 0.06386801737041177 1 -634 0 -5.19334 0.0047492543 0.0068680476530536229 0 -635 0 -4.540021 0.008633798 0.012510020082825467 0 -636 1 9.007486 0.99956155 0.0006326909022333528 1 -637 0 -2.4041815 0.058601014 0.087121796299934065 0 -638 0 -4.6241584 0.007995274 0.011581101010143969 0 -639 0 -3.3673954 0.024999332 0.036524887125210473 0 -640 0 -3.920342 0.015175682 0.022061707904714886 0 -641 0 -4.9072876 0.006171662 0.0089314152845014639 0 -642 0 -4.9072876 0.006171662 0.0089314152845014639 0 -643 0 -6.331703 0.0016699884 0.002411297942606464 0 -644 0 -5.7625213 0.002817351 0.0040703146280809182 0 -645 0 -4.9072876 0.006171662 0.0089314152845014639 0 -646 0 -5.834298 0.0026376322 0.0038103262134336524 0 -647 0 -5.627887 0.00318803 0.0046067022700437095 0 -648 1 9.190195 0.99962944 0.00053470714742229219 1 -649 0 -4.9072876 0.006171662 0.0089314152845014639 0 -650 0 -3.939826 0.014909853 0.021672341535773916 0 -651 0 -5.1132717 0.0051107947 0.007392224339633548 0 -652 0 -3.5992842 0.0202901 0.029573474708830459 0 -653 0 -4.338106 0.010379729 0.015053042459133558 0 -654 0 -4.621235 0.008016654 0.011612195307125778 0 -655 0 -4.9072876 0.006171662 0.0089314152845014639 0 -656 0 -3.7689247 0.017406758 0.025333778099074687 0 -657 0 -0.46601295 0.2705457 0.45511048083845856 0 -658 1 8.120775 0.9990085 0.0014311753376832121 1 -659 0 -6.331703 0.0016699884 0.002411297942606464 0 -660 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -661 0 -4.052054 0.013465655 0.019558817434379178 0 -662 0 -5.5923867 0.0032936204 0.004759532206015593 0 -663 0 -5.5923867 0.0032936204 0.004759532206015593 0 -664 0 -4.203471 0.011733688 0.01702823259872497 0 -665 0 -6.331703 0.0016699884 0.002411297942606464 0 -666 0 -2.9374223 0.03669785 0.053939708586813642 0 -667 0 -4.621235 0.008016654 0.011612195307125778 0 -668 1 2.1150274 0.7997754 0.32233314938902585 1 -669 1 7.066201 0.9973858 0.0037764318699177366 1 -670 1 5.4009876 0.98799986 0.017417262659521413 1 -671 0 -4.0500174 0.01349059 0.019595282853318546 0 -672 0 -5.1904163 0.0047619957 0.0068865174780414111 0 -673 0 -3.4082727 0.024098057 0.035191899852455667 0 -674 0 -6.0456505 0.0021721658 0.0031371812770679881 0 -675 0 -3.9892325 0.014256168 0.020715316992422194 0 -676 0 -5.8559704 0.00258565 0.0037351353289297126 0 -677 0 -4.054977 0.013429944 0.019506595612928591 0 -678 0 -6.331703 0.0016699884 0.002411297942606464 0 -679 0 -5.7625213 0.002817351 0.0040703146280809182 0 -680 1 14.850252 0.999998 2.9237080272005804E-06 1 -681 1 8.980978 0.9995507 0.00064834827339000788 1 -682 0 -3.3132315 0.026244283 0.038368201095627046 0 -683 0 -6.331703 0.0016699884 0.002411297942606464 0 -684 0 -6.331703 0.0016699884 0.002411297942606464 0 -685 0 -6.331703 0.0016699884 0.002411297942606464 0 -686 0 -6.331703 0.0016699884 0.002411297942606464 0 -687 0 -4.5680065 0.008415992 0.012193091213284017 0 -688 0 -4.6241584 0.007995274 0.011581101010143969 0 -689 0 -4.4540114 0.009338773 0.013536306372392479 0 -690 0 -5.627887 0.00318803 0.0046067022700437095 0 -691 1 4.4992046 0.9728897 0.039651812404386706 1 -692 0 -5.19334 0.0047492543 0.0068680476530536229 0 -693 0 -4.6390457 0.007887273 0.011424041506677902 0 -694 0 -4.998421 0.005677692 0.0082145201627354687 0 -695 0 -5.7625213 0.002817351 0.0040703146280809182 0 -696 1 6.3037586 0.9947386 0.0076106653397057331 1 -697 1 3.8583202 0.9521294 0.070770401478791711 1 -698 1 4.4452257 0.9715474 0.041643674121419214 1 +618 0 -4.556242 0.00918188 0.013307842744969949 0 +619 0 -4.0081143 0.015174806 0.022060425445491177 0 +620 0 -5.10437 0.005542395 0.0080182267005007406 0 +621 0 -0.41670036 0.30127206 0.51719727367063018 0 +622 0 -1.9880915 0.0912099 0.13798098144151136 0 +623 0 -6.457423 0.0015860328 0.0022899781237268467 0 +624 0 -3.455483 0.0250826 0.036648102437293467 0 +625 0 -3.6625319 0.02079054 0.030310598654120183 0 +626 1 4.185855 0.9685779 0.046059977621398726 1 +627 0 -4.532421 0.00938511 0.013603788328355914 0 +628 0 -5.9092956 0.0026344028 0.0038056549166740083 0 +629 0 -4.81304 0.0072497097 0.010497217242063378 0 +630 0 -3.0560293 0.035928845 0.052788464415381341 0 +631 0 -4.0081143 0.015174806 0.022060425445491177 0 +632 0 -6.457423 0.0015860328 0.0022899781237268467 0 +633 1 3.7827978 0.95497304 0.06646808708515968 1 +634 0 -5.361168 0.004372712 0.0063223231877638552 0 +635 0 -4.5400214 0.009319788 0.013508658860235377 0 +636 1 8.132348 0.9991667 0.0012026598711080015 1 +637 0 -2.495535 0.058984704 0.087709921531216192 0 +638 0 -4.81304 0.0072497097 0.010497217242063378 0 +639 0 -3.5227346 0.023601409 0.034457881730402082 0 +640 0 -4.070862 0.0143290255 0.020821952043458752 0 +641 0 -5.10437 0.005542395 0.0080182267005007406 0 +642 0 -5.10437 0.005542395 0.0080182267005007406 0 +643 0 -6.457423 0.0015860328 0.0022899781237268467 0 +644 0 -5.9092956 0.0026344028 0.0038056549166740083 0 +645 0 -5.10437 0.005542395 0.0080182267005007406 0 +646 0 -5.978259 0.0024715506 0.0035701074650224761 0 +647 0 -5.715246 0.003152347 0.0045550586337529753 0 +648 1 8.424798 0.99936455 0.00091704491714961699 1 +649 0 -5.10437 0.005542395 0.0080182267005007406 0 +650 0 -4.1432714 0.01341063 0.019478351346481192 0 +651 0 -5.1608734 0.0052608596 0.007609850855068395 0 +652 0 -3.8581696 0.017400013 0.025323875300941854 0 +653 0 -4.556242 0.00918188 0.013307842744969949 0 +654 0 -4.8398457 0.007072917 0.010240319857589488 0 +655 0 -5.10437 0.005542395 0.0080182267005007406 0 +656 0 -4.0081143 0.015174806 0.022060425445491177 0 +657 0 -0.466012 0.29173103 0.49763075667347878 0 +658 1 7.574725 0.998603 0.0020168738138479237 1 +659 0 -6.457423 0.0015860328 0.0022899781237268467 0 +660 0 -6.200626 0.0020117955 0.0029053308785867587 0 +661 0 -4.289057 0.011734198 0.017028976282173813 0 +662 0 -5.5923867 0.0035315522 0.0051039705605183091 0 +663 0 -5.5923867 0.0035315522 0.0051039705605183091 0 +664 0 -4.362192 0.010972952 0.015918117738024398 0 +665 0 -6.457423 0.0015860328 0.0022899781237268467 0 +666 0 -3.1347342 0.033483833 0.049134230274882645 0 +667 0 -4.8398457 0.007072917 0.010240319857589488 0 +668 1 2.1150274 0.81866264 0.28865902997325171 1 +669 1 6.8052893 0.99715185 0.0041148721631914234 1 +670 1 4.8467693 0.9827309 0.02513163883784484 1 +671 0 -4.1789083 0.012980201 0.01884907076662112 0 +672 0 -5.3906345 0.0042553055 0.0061522072207014699 0 +673 0 -3.6740208 0.02057467 0.029992587648424631 0 +674 0 -6.200626 0.0020117955 0.0029053308785867587 0 +675 0 -3.989233 0.015438799 0.02244720678817614 0 +676 0 -5.990636 0.002443405 0.0035294020597964296 0 +677 0 -4.2649117 0.0119968 0.01741238050521976 0 +678 0 -6.457423 0.0015860328 0.0022899781237268467 0 +679 0 -5.9092956 0.0026344028 0.0038056549166740083 0 +680 1 14.0086565 0.9999964 5.1594887517745861E-06 1 +681 1 8.4504 0.99937946 0.0008955335767871287 1 +682 0 -3.6013718 0.021977525 0.032060476434216285 0 +683 0 -6.457423 0.0015860328 0.0022899781237268467 0 +684 0 -6.457423 0.0015860328 0.0022899781237268467 0 +685 0 -6.457423 0.0015860328 0.0022899781237268467 0 +686 0 -6.457423 0.0015860328 0.0022899781237268467 0 +687 0 -4.8044806 0.007307079 0.010580590717855092 0 +688 0 -4.81304 0.0072497097 0.010497217242063378 0 +689 0 -4.4979525 0.009687093 0.014043652597884091 0 +690 0 -5.715246 0.003152347 0.0045550586337529753 0 +691 1 4.4992037 0.9763163 0.034579516892942586 1 +692 0 -5.361168 0.004372712 0.0063223231877638552 0 +693 0 -4.8244457 0.007173954 0.010387131031373918 0 +694 0 -5.202798 0.0050612013 0.0073203104438950544 0 +695 0 -5.9092956 0.0026344028 0.0038056549166740083 0 +696 1 6.3037596 0.99547225 0.0065469918272239849 1 +697 1 3.1989088 0.9250327 0.11242376751569673 1 +698 1 4.013835 0.9633409 0.053881708362724622 1 diff --git a/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Default-CV-breast-cancer.txt b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Default-CV-breast-cancer.txt new file mode 100644 index 0000000000..d3229b5a8a --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Default-CV-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +5 1 7.0572085 0.99996483 5.0735774115656486E-05 1 +6 0 -0.56973577 0.32680443 0.57090241040548417 0 +8 0 -3.265584 0.009920232 0.014383331043645656 0 +9 0 -2.8438797 0.018053673 0.0262839259348677 0 +10 0 -3.4260702 0.007890133 0.01142820055619111 0 +11 0 -3.2730165 0.009815701 0.014231022120774742 0 +18 1 4.1297426 0.9976288 0.0034249691446541306 1 +20 1 3.6110172 0.99500996 0.0072171296745399023 1 +21 1 3.7079864 0.9956572 0.0062789720030828463 1 +25 1 0.8509629 0.78957224 0.34085682884364138 1 +28 0 -3.2730165 0.009815701 0.014231022120774742 0 +31 0 -3.1337807 0.0119679235 0.017370215154790519 0 +32 1 3.221815 0.9912946 0.012614190719505857 1 +35 0 -3.2730165 0.009815701 0.014231022120774742 0 +37 0 -2.1029625 0.05070596 0.075073066649086931 0 +40 0 ? ? ? 0 +41 1 0.42940283 0.671623 0.57427647621967726 1 +44 1 4.518715 0.99864405 0.0019575441244945968 1 +45 0 -3.4117298 0.008053371 0.011665595031047894 0 +46 1 3.4790392 0.99397224 0.0087225318959698473 1 +48 0 -2.6620011 0.023330595 0.034057791625520743 0 +50 1 0.44066358 0.6751879 0.56663907548228165 1 +51 1 -0.21027565 0.448867 1.1556400813104404 0 +52 1 2.103271 0.95791155 0.062035644381648124 1 +54 1 3.5159762 0.99428254 0.0082722165268463033 1 +56 1 3.9441156 0.99690473 0.0044724548318091735 1 +60 1 0.9666352 0.8159061 0.29352495512337407 1 +63 1 -0.13535523 0.47566786 1.0719735341494714 0 +64 0 -3.4788275 0.007317354 0.010595523334913478 0 +66 0 -2.900283 0.016669292 0.024251398007332337 0 +68 1 4.283942 0.9980999 0.0027438390509451197 1 +69 0 -3.2245307 0.010517775 0.015254303988182968 0 +70 0 -2.5950685 0.025629777 0.037458050777115802 0 +71 1 3.5624905 0.99465084 0.007737919692094284 1 +72 0 -1.587163 0.10090509 0.1534546786625488 0 +73 1 4.7288055 0.99899757 0.0014469274547371505 1 +74 1 0.4975586 0.6928849 0.5293123330101398 1 +76 0 -2.8515797 0.01785823 0.025996804845523382 0 +77 0 -2.3885822 0.03419739 0.050199732728753617 0 +79 0 -3.1797595 0.011210072 0.016264046920404552 0 +82 0 -2.6207623 0.024722181 0.036114849743136321 0 +88 0 -2.900283 0.016669292 0.024251398007332337 0 +90 0 -3.3328133 0.009013457 0.013062628597518949 0 +91 0 -3.1149797 0.012292199 0.017843790083392992 0 +92 0 -2.900283 0.016669292 0.024251398007332337 0 +93 0 -3.4788275 0.007317354 0.010595523334913478 0 +95 0 -3.3328133 0.009013457 0.013062628597518949 0 +96 0 -3.3179266 0.009206889 0.013344257636588433 0 +97 0 -2.7835343 0.019659806 0.028645620182097348 0 +98 1 4.165916 0.99774885 0.0032513814242770777 1 +99 1 5.3154254 0.9995689 0.00062210936850165717 1 +100 1 2.4310963 0.97332186 0.039011145041019712 1 +102 0 -2.7532477 0.020517876 0.029908932550911323 0 +104 1 5.4333305 0.9996362 0.00052498655830247973 1 +105 1 -0.15268373 0.46945086 1.0909539401833979 0 +106 1 5.426038 0.9996323 0.00053057803961755552 1 +108 0 -3.2805295 0.009711145 0.014078692642497696 0 +109 1 3.3400679 0.9926472 0.010647081134485804 1 +111 1 2.1333807 0.95962465 0.059457880954894356 1 +112 1 4.0789456 0.9974494 0.0036844415698049165 1 +113 1 5.289748 0.99955267 0.00064550928688674445 1 +115 0 -2.3835738 0.034436304 0.050556661505383516 0 +117 1 4.4668274 0.99853903 0.0021092745459332951 1 +120 0 -2.9493718 0.015549762 0.022609812535845208 0 +121 0 -2.4117963 0.03311076 0.04857746089833543 0 +122 1 5.8156424 0.99979013 0.00030280723272274124 1 +123 1 2.3205795 0.9688648 0.045632737891349161 1 +125 0 -3.4788275 0.007317354 0.010595523334913478 0 +128 1 2.6282647 0.97978055 0.029469435449918441 1 +129 0 -2.7184992 0.02154758 0.03142639763558145 0 +131 0 -3.1337807 0.0119679235 0.017370215154790519 0 +132 1 5.7164726 0.99975795 0.00034925304341121586 1 +133 0 -3.039434 0.0136849675 0.019879572841258564 0 +137 0 -3.2418132 0.010262002 0.014881427628699428 0 +138 0 -2.8588524 0.01767554 0.025728471293859974 0 +141 0 -3.4372978 0.007764623 0.011245699174135494 0 +144 0 -3.2730165 0.009815701 0.014231022120774742 0 +145 0 ? ? ? 0 +147 0 -3.0207539 0.014052695 0.020417552098652373 0 +150 0 -3.4260702 0.007890133 0.01142820055619111 0 +151 1 2.217757 0.9640765 0.052780436794473695 1 +152 1 5.1777344 0.99947447 0.00075838481234789161 1 +154 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +156 0 -3.0233922 0.014000173 0.020340701014534553 0 +161 0 -2.8170805 0.018750407 0.027307944101129199 0 +164 0 ? ? ? 0 +167 1 2.9291844 0.98679465 0.019178200124059363 1 +169 0 -3.601779 0.006137724 0.0088821500896236558 0 +171 0 -3.3328133 0.009013457 0.013062628597518949 0 +173 1 7.947592 0.9999902 1.4102646298686894E-05 1 +174 1 3.2498922 0.9916365 0.012116698660061622 1 +176 0 -3.1337807 0.0119679235 0.017370215154790519 0 +177 1 3.3192258 0.9924249 0.010970154193342282 1 +179 1 1.378361 0.88909316 0.16959349997611742 1 +180 0 -3.4260702 0.007890133 0.01142820055619111 0 +181 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +183 1 5.419508 0.99962884 0.00053556737970243554 1 +187 1 5.8260784 0.99979323 0.00029833475210104223 1 +188 1 4.9982443 0.9993196 0.00098192506222582999 1 +189 0 -2.6896749 0.022439936 0.03274274682140442 0 +191 1 6.3130674 0.9998974 0.00014799866201853666 1 +192 0 -2.889139 0.016934283 0.024640232432541796 0 +196 0 2.9049587 0.9863325 6.1931042955602695 1 +198 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +199 0 -3.1226366 0.012159102 0.017649394845221815 0 +201 1 5.4355135 0.9996373 0.00052335212939580867 1 +202 0 -3.3328133 0.009013457 0.013062628597518949 0 +204 0 -3.3328133 0.009013457 0.013062628597518949 0 +205 1 7.0590024 0.99996495 5.0563785426763426E-05 1 +206 1 3.8042984 0.9962173 0.0054676151631998047 1 +207 0 -3.4260702 0.007890133 0.01142820055619111 0 +209 0 -2.88948 0.01692611 0.024628237800566447 0 +210 1 7.616646 0.99998426 2.2701888528575377E-05 1 +211 1 5.476496 0.99965805 0.00049341660116784 1 +212 0 -3.3328133 0.009013457 0.013062628597518949 0 +216 0 -3.4788275 0.007317354 0.010595523334913478 0 +218 1 4.5168786 0.9986405 0.0019627106187464437 1 +219 0 -2.360191 0.03557316 0.052256290809923854 0 +223 1 2.745268 0.98286086 0.024940896191451886 1 +226 1 4.9640284 0.9992853 0.0010314906403046983 1 +228 0 -3.4260702 0.007890133 0.01142820055619111 0 +233 1 3.1638048 0.99054366 0.013707523882617778 1 +237 1 3.3809423 0.9930644 0.010040810521379973 1 +239 1 2.5812135 0.9783943 0.031512052384344527 1 +240 0 -1.7316827 0.083536625 0.12585086783696176 0 +241 0 -2.8964546 0.016759865 0.024384288407907153 0 +242 0 -3.1337807 0.0119679235 0.017370215154790519 0 +244 0 -3.3328133 0.009013457 0.013062628597518949 0 +246 1 6.6648927 0.9999382 8.9175760478377495E-05 1 +247 1 1.5379701 0.90980536 0.13637016481621997 1 +248 0 -2.3839004 0.034420673 0.050533306192156652 0 +249 0 ? ? ? 0 +250 0 -3.1919966 0.0110164955 0.015981636806704738 0 +252 0 2.360736 0.97056204 5.0861785306924734 1 +254 1 4.4944267 0.9985959 0.0020271211335618432 1 +257 0 -3.1226366 0.012159102 0.017649394845221815 0 +258 0 -3.017032 0.014127118 0.020526456121260601 0 +259 0 2.3594117 0.9705075 5.083508193833687 1 +260 1 4.7412558 0.99901533 0.0014212765543103185 1 +262 1 5.535843 0.99968606 0.00045298741967439286 1 +267 1 1.7871771 0.93523186 0.096604009859263623 1 +268 1 3.395525 0.99320745 0.0098330049477988995 1 +269 0 -3.3328133 0.009013457 0.013062628597518949 0 +271 0 -2.7835343 0.019659806 0.028645620182097348 0 +272 1 1.7871771 0.93523186 0.096604009859263623 1 +275 0 ? ? ? 0 +276 0 -3.1226366 0.012159102 0.017649394845221815 0 +277 0 -3.4788275 0.007317354 0.010595523334913478 0 +278 0 -3.3328133 0.009013457 0.013062628597518949 0 +279 1 2.5009675 0.9758128 0.035323697956253994 1 +280 0 -3.017032 0.014127118 0.020526456121260601 0 +283 1 2.839066 0.98499304 0.021814563600203427 1 +284 1 4.045544 0.9973241 0.0038656689278478121 1 +285 1 7.694996 0.999986 2.0208103005084558E-05 1 +288 1 0.650548 0.7376639 0.43896441221797339 1 +290 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +291 0 -3.3328133 0.009013457 0.013062628597518949 0 +293 1 2.8659723 0.98555493 0.02099180743545831 1 +296 0 0.84631276 0.78845793 2.2409834905598212 1 +297 0 ? ? ? 0 +299 1 3.3291118 0.9925311 0.010815756272127022 1 +300 1 3.6169615 0.9950523 0.007155770949232791 1 +301 0 -3.3328133 0.009013457 0.013062628597518949 0 +303 0 -3.3328133 0.009013457 0.013062628597518949 0 +304 1 3.2374554 0.9914867 0.012334633881181759 1 +308 1 3.36895 0.9929445 0.010215043928391239 1 +309 0 -1.5727417 0.10280404 0.15650497464376914 0 +311 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +312 1 1.5964963 0.916484 0.1258184031857312 1 +314 0 -3.5903516 0.0062388843 0.009029002258266916 0 +316 1 2.3232007 0.9689784 0.045463581309268399 1 +317 1 5.1857476 0.9994805 0.00074969514792004902 1 +319 0 1.087976 0.84070957 2.6502684744706926 1 +321 0 ? ? ? 0 +323 1 3.347725 0.99272716 0.010530830655206472 1 +327 0 -3.4788275 0.007317354 0.010595523334913478 0 +328 1 2.392449 0.97183865 0.041211280760177167 1 +329 1 3.6536255 0.9953054 0.0067887965489780632 1 +331 0 -2.764847 0.020185 0.029418715877423222 0 +332 0 -2.3301463 0.037087068 0.054522741465995221 0 +333 1 3.065022 0.98911446 0.015790611190946886 1 +336 1 3.449659 0.99371344 0.0090982194011089285 1 +338 0 -3.5903516 0.0062388843 0.009029002258266916 0 +343 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +344 1 5.3123817 0.999567 0.00062486228037730135 1 +346 0 -2.2415557 0.04191981 0.061781680518225482 0 +347 0 -3.4701545 0.0074085975 0.010728136357334885 0 +348 1 0.009417295 0.52771837 0.92215990307325479 1 +349 1 1.9216316 0.94601345 0.080067398628708528 1 +350 0 -2.7972674 0.019282404 0.028090332536416556 0 +352 0 0.48690295 0.68961143 1.6878526862994818 1 +353 1 5.684911 0.9997467 0.00036550943041529892 1 +354 0 -3.4788275 0.007317354 0.010595523334913478 0 +355 0 -2.8206 0.018657422 0.027171238007066462 0 +358 1 3.429872 0.993533 0.0093601856431956044 1 +360 1 8.039464 0.9999915 1.2296811942483466E-05 1 +361 1 4.0018487 0.99715084 0.0041163381919338436 1 +366 1 7.6356616 0.99998474 2.2013947263955502E-05 1 +368 0 -3.1859481 0.011111759 0.016120610060422667 0 +370 0 -2.4288907 0.032331984 0.047415917766564633 0 +371 0 -3.1859481 0.011111759 0.016120610060422667 0 +373 0 -2.7758775 0.019873358 0.028959923684981098 0 +376 0 -3.4788275 0.007317354 0.010595523334913478 0 +377 0 -3.5903516 0.0062388843 0.009029002258266916 0 +378 0 -2.5109992 0.028831733 0.042206813278669819 0 +379 0 -1.6979227 0.08733312 0.13183971754063689 0 +381 1 4.5339613 0.99867344 0.0019150934639560221 1 +383 0 -3.4372978 0.007764623 0.011245699174135494 0 +384 0 -3.4372978 0.007764623 0.011245699174135494 0 +387 0 -1.8698577 0.06951699 0.1039482817545123 0 +388 0 -3.164734 0.011452365 0.01661760793709334 0 +389 0 -2.384867 0.034374464 0.0504642666238184 0 +391 1 5.278062 0.9995451 0.00065643511405119592 1 +392 0 -3.1226366 0.012159102 0.017649394845221815 0 +395 0 -3.1226366 0.012159102 0.017649394845221815 0 +396 0 -3.017032 0.014127118 0.020526456121260601 0 +398 0 -2.8773108 0.017220069 0.025059697242952091 0 +399 0 -2.983367 0.014818167 0.021538070923183442 0 +404 0 -3.035563 0.013760382 0.019989887173534629 0 +406 0 -2.6738753 0.022944283 0.033487259692482632 0 +409 0 -2.9756012 0.014982245 0.021778365324536704 0 +413 0 -2.636499 0.02418186 0.035315791713226907 0 +414 1 3.5129192 0.9942575 0.0083085410219168807 1 +415 0 -0.4454751 0.3673024 0.6604119441020434 0 +416 1 5.0948877 0.9994079 0.00085449082917629041 1 +418 0 -1.7674735 0.07967611 0.11978641066696791 0 +419 0 -2.7845545 0.019631524 0.028603999634855711 0 +422 0 -1.93198 0.06395076 0.095343674769021142 0 +423 0 -2.5950685 0.025629777 0.037458050777115802 0 +428 0 -3.4788275 0.007317354 0.010595523334913478 0 +429 0 -3.2730165 0.009815701 0.014231022120774742 0 +430 0 -2.9450116 0.015646132 0.02275104735908454 0 +434 0 3.3375685 0.9926208 7.0823249043326983 1 +436 1 2.481813 0.9751534 0.036298929694796755 1 +439 0 -3.00913 0.014286414 0.020759583962433738 0 +440 1 4.7098246 0.9989698 0.0014870401797071949 1 +441 0 -1.2440159 0.15534686 0.24356907239421263 0 +442 0 -2.68195 0.022685155 0.033104689547068904 0 +449 1 5.942175 0.99982506 0.0002524066187454977 1 +450 0 -2.8853104 0.017026272 0.024775236323585029 0 +451 0 -3.00913 0.014286414 0.020759583962433738 0 +452 0 -3.001718 0.014437441 0.020980644716606881 0 +453 1 4.8444147 0.9991511 0.0012252085636515727 1 +454 0 -2.9710684 0.015078842 0.02191985163250678 0 +455 1 -0.41451907 0.37771752 1.4046203735720693 0 +456 1 5.1211615 0.9994299 0.00082274158011057654 1 +457 1 4.6062374 0.99880433 0.0017260180672053755 1 +464 0 -3.1254716 0.012110182 0.017577951873598927 0 +465 1 5.048981 0.9993675 0.00091282866916632312 1 +466 1 4.630232 0.9988449 0.0016673890635841846 1 +467 1 4.1375737 0.99765533 0.0033866125627017287 1 +474 0 -3.00913 0.014286414 0.020759583962433738 0 +480 0 -3.1006896 0.0125444615 0.018212304550370879 0 +482 1 7.239645 0.999973 3.8954596327452357E-05 1 +483 1 5.804901 0.99978685 0.00030753775616068136 1 +484 0 -2.866447 0.017486723 0.025451192013999709 0 +487 1 6.835632 0.99995166 6.9740650547897242E-05 1 +489 1 -0.9796454 0.21203241 2.2376433084084262 0 +492 0 -3.1118338 0.012347301 0.017924276606752516 0 +493 1 6.1447577 0.9998693 0.00018859130260575588 1 +495 0 -3.2285826 0.010457248 0.015166056993009934 0 +497 0 -2.979243 0.014905077 0.021665347213202568 0 +501 0 -3.005888 0.014352282 0.020855991589474016 0 +502 0 -2.8926263 0.016850917 0.024517893505160124 0 +504 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +507 0 -2.6673017 0.023157371 0.033801934892946583 0 +510 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +513 0 -3.2285826 0.010457248 0.015166056993009934 0 +514 1 5.730319 0.9997627 0.00034237208821625534 1 +517 0 -3.5903516 0.0062388843 0.009029002258266916 0 +519 1 3.8362005 0.99638647 0.0052226659316945189 1 +520 0 -3.4750462 0.007356997 0.010653139036342341 0 +521 0 -3.0694122 0.013114572 0.019045488860345496 0 +522 1 1.8001244 0.93635166 0.094877643704389392 1 +523 1 4.2548237 0.99801874 0.0028611869695542697 1 +527 0 -2.900283 0.016669292 0.024251398007332337 0 +528 0 -2.4706626 0.030502755 0.044691296515759281 0 +529 0 -3.1118338 0.012347301 0.017924276606752516 0 +531 0 -2.6738753 0.022944283 0.033487259692482632 0 +532 0 -3.4260702 0.007890133 0.01142820055619111 0 +533 0 -3.1226366 0.012159102 0.017649394845221815 0 +534 0 -3.2730165 0.009815701 0.014231022120774742 0 +535 0 -2.7646904 0.02018946 0.029425284390774787 0 +538 0 -3.005888 0.014352282 0.020855991589474016 0 +539 0 -2.7723901 0.019971373 0.029104202738105003 0 +540 0 -2.6079202 0.025171826 0.036780146667765946 0 +541 0 -3.2418132 0.010262002 0.014881427628699428 0 +544 0 -2.7833471 0.019665 0.028653262441506629 0 +546 1 6.9030447 0.99995613 6.3291003784773738E-05 1 +547 0 -3.5305548 0.0067958706 0.0098378351405619836 0 +548 0 -3.3341873 0.008995808 0.013036934393021076 0 +549 1 3.0349069 0.9886377 0.016486193421270071 1 +557 0 -2.7972674 0.019282404 0.028090332536416556 0 +558 0 -3.2730165 0.009815701 0.014231022120774742 0 +559 0 -2.889139 0.016934283 0.024640232432541796 0 +560 0 -2.7835343 0.019659806 0.028645620182097348 0 +561 0 -2.7835343 0.019659806 0.028645620182097348 0 +563 0 -3.1226366 0.012159102 0.017649394845221815 0 +565 1 6.7825165 0.9999478 7.533036773246141E-05 1 +566 0 -2.8699965 0.017399155 0.025322614551901095 0 +569 1 5.348246 0.9995888 0.00059337616434588982 1 +577 0 -3.4788275 0.007317354 0.010595523334913478 0 +578 0 -3.4788275 0.007317354 0.010595523334913478 0 +581 1 5.0124683 0.9993334 0.00096204767857963188 1 +582 1 5.0096254 0.99933064 0.00096600592361462709 1 +584 0 -2.2400014 0.042009756 0.061917130950783161 0 +586 1 7.861235 0.999989 1.5908482915272255E-05 1 +590 1 2.2271297 0.96454084 0.052085770809847762 1 +593 0 -2.866447 0.017486723 0.025451192013999709 0 +594 1 3.2574682 0.9917265 0.011985762568813323 1 +600 0 -3.1226366 0.012159102 0.017649394845221815 0 +602 0 -3.005888 0.014352282 0.020855991589474016 0 +604 1 2.9559016 0.98728657 0.018459195843063887 1 +606 0 -2.998231 0.014509038 0.021085454705547534 0 +607 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +609 0 -3.00913 0.014286414 0.020759583962433738 0 +612 1 8.095181 0.99999213 1.135089961038231E-05 1 +613 0 -2.8651934 0.017517753 0.025496756032881891 0 +614 0 -3.3662734 0.008593262 0.012451031160595608 0 +617 0 ? ? ? 0 +618 0 -3.005888 0.014352282 0.020855991589474016 0 +619 0 -2.889139 0.016934283 0.024640232432541796 0 +621 0 -0.85922766 0.24243373 0.40055599131478326 0 +622 0 -2.2535048 0.04123443 0.060749993133213671 0 +624 0 -2.8818235 0.017110478 0.02489883011380966 0 +627 0 -2.2639332 0.040645044 0.059863392172501514 0 +629 0 -3.1254716 0.012110182 0.017577951873598927 0 +633 1 2.2230194 0.96433794 0.052389278189625965 1 +634 0 -3.2418132 0.010262002 0.014881427628699428 0 +638 0 -3.1254716 0.012110182 0.017577951873598927 0 +639 0 -2.7972674 0.019282404 0.028090332536416556 0 +641 0 -3.1226366 0.012159102 0.017649394845221815 0 +642 0 -3.1226366 0.012159102 0.017649394845221815 0 +644 0 -3.4372978 0.007764623 0.011245699174135494 0 +645 0 -3.1226366 0.012159102 0.017649394845221815 0 +649 0 -3.1226366 0.012159102 0.017649394845221815 0 +652 0 -2.8059363 0.01904784 0.027745314802772762 0 +653 0 -3.005888 0.014352282 0.020855991589474016 0 +654 0 -3.017032 0.014127118 0.020526456121260601 0 +656 0 -2.889139 0.016934283 0.024640232432541796 0 +657 0 -0.18228126 0.45885497 0.8859128072339173 0 +660 0 -3.4788275 0.007317354 0.010595523334913478 0 +661 0 -2.900283 0.016669292 0.024251398007332337 0 +665 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +668 1 1.1573684 0.8536355 0.22830793735552202 1 +670 1 4.2615523 0.9980378 0.0028336153819191402 1 +678 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +679 0 -3.4372978 0.007764623 0.011245699174135494 0 +680 1 8.287129 0.99999404 8.5991581706542968E-06 1 +681 1 5.993926 0.99983764 0.00023425938852449468 1 +682 0 -2.7003317 0.022105886 0.032249835231867321 0 +683 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +685 0 -3.6054602 0.0061054854 0.0088353530585513502 0 +688 0 -3.1254716 0.012110182 0.017577951873598927 0 +689 0 -3.2345 0.010369475 0.015038094181500315 0 +691 1 3.3198962 0.9924322 0.010959583213904138 1 +692 0 -3.2418132 0.010262002 0.014881427628699428 0 +693 0 -3.1039493 0.012486473 0.018127584079738208 0 +694 0 -2.8741245 0.017297858 0.025173894917894488 0 +696 1 4.167007 0.99775237 0.00324629649809509 1 +697 1 2.6165223 0.97944295 0.029966627189265894 1 +698 1 3.2709064 0.9918837 0.011757129822144028 1 +0 0 -2.1858046 0.019522207 0.028443139723819993 0 +1 0 2.4643464 0.8795063 3.0529702573901116 1 +2 0 -2.3502047 0.015902877 0.023127388909652609 0 +3 0 2.427892 0.8745144 2.9944062875580588 1 +4 0 -2.082891 0.022186846 0.032369280924918455 0 +7 0 -2.638155 0.01108701 0.016084504095792203 0 +12 1 0.13348675 0.27454057 1.8649087192175371 1 +13 0 -2.494972 0.01326812 0.019269973170525343 0 +14 1 5.299982 0.99627197 0.0053884638761153361 1 +15 1 0.7744734 0.46061882 1.1183547250072077 1 +16 0 -2.4918776 0.013319657 0.019345327864414775 0 +17 0 -2.400572 0.014932421 0.021705393169323427 0 +19 0 -1.959885 0.02584031 0.037769808679167967 0 +22 0 -2.7270536 0.009915355 0.014376224049015385 0 +23 1 ? ? ? 0 +24 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +26 0 -2.5670438 0.012121931 0.017595109452200379 0 +27 0 -2.280314 0.017352993 0.025254839289623091 0 +29 0 -2.9168253 0.007808857 0.011310016307330938 0 +30 0 -2.7434447 0.009713116 0.014081563613944885 0 +33 0 -2.697853 0.01028602 0.014916437796865623 0 +34 0 -2.5725198 0.012038953 0.017473933576498117 0 +36 1 5.3707566 0.99659127 0.0049261586606575156 1 +38 1 3.8173842 0.97600687 0.035036798748799364 1 +39 1 1.7423136 0.74478805 0.42509816834409625 1 +42 1 5.30029 0.9962734 0.0053863923631196743 1 +43 1 0.43130207 0.35581547 1.4907988571250606 1 +47 0 -3.1699133 0.005674961 0.0082105575200780885 0 +49 1 4.063138 0.98232436 0.025728616279702864 1 +53 1 3.6499045 0.9704895 0.043215487885653277 1 +55 1 3.0962358 0.9421367 0.085991683453461656 1 +57 1 0.31318235 0.3222287 1.6338430986952079 1 +58 1 1.486213 0.67827094 0.56006642040098542 1 +59 1 0.84864306 0.4840856 1.0466659464721533 1 +61 0 -2.9074705 0.007901425 0.011444620183650155 0 +62 1 4.689613 0.9919433 0.01167043745866826 1 +65 1 1.4220569 0.66024333 0.59893026779783309 1 +67 1 3.0339043 0.93766826 0.092850490546770872 1 +75 0 -2.2670527 0.017642453 0.025679879972587048 0 +78 0 -2.0084546 0.024332527 0.035538562774444914 0 +80 0 -2.03422 0.023567846 0.03440829097294839 0 +81 0 -2.3108268 0.016704578 0.024303168767298499 0 +83 0 -1.8550333 0.029411687 0.043068605771028841 0 +84 1 5.2740955 0.9961479 0.0055681789505025291 1 +85 1 3.9099104 0.97860944 0.031194891325044546 1 +86 1 1.9654632 0.79483765 0.33126787626487048 1 +87 1 4.4889183 0.98962927 0.015039925710551912 1 +89 0 -2.795878 0.009093219 0.013178752422657699 0 +94 0 -2.8278565 0.008734563 0.012656667304014846 0 +101 1 0.26796222 0.30981937 1.6905007478467695 1 +103 1 0.2317195 0.30006656 1.7366455412926651 1 +107 1 4.468525 0.9893601 0.015432385453245953 1 +110 0 -1.4909216 0.04590447 0.067794367893782667 0 +114 0 -1.4255109 0.04968219 0.073518025658764419 0 +116 0 -0.30933428 0.17741625 0.28176552315143077 0 +118 0 -2.611174 0.01146897 0.016641842325176619 0 +119 0 -2.1002054 0.02171489 0.03167311202362464 0 +124 1 4.1922474 0.9849567 0.021867818470544111 1 +126 1 4.365718 0.9878945 0.017571150102842548 1 +127 0 -2.6042523 0.011569039 0.016787893374491106 0 +130 0 -1.920393 0.027132994 0.039685496735227151 0 +134 0 -2.7086492 0.010147406 0.014714395385468908 0 +135 0 -1.4709823 0.047026105 0.069491400578601079 0 +136 0 -2.4918776 0.013319657 0.019345327864414775 0 +139 0 ? ? ? 0 +140 0 -2.8614266 0.008373139 0.012130744104216924 0 +142 1 2.578139 0.8939979 0.16165664003695771 1 +143 0 -2.1831431 0.019586993 0.028538471086822469 0 +146 1 1.1117849 0.56719834 0.81807479444774811 1 +148 0 -0.79389584 0.104406506 0.15908404676927829 0 +149 1 6.315874 0.99897087 0.0014854907404379277 1 +153 0 -2.1021397 0.02166278 0.031596264691020635 0 +155 1 2.581787 0.89443606 0.16094973974671178 1 +157 0 -2.8278565 0.008734563 0.012656667304014846 0 +158 0 ? ? ? 0 +159 1 7.2913313 0.9997015 0.00043070884426719654 1 +160 1 5.6563535 0.9976256 0.0034296237206113307 1 +162 0 -2.6042523 0.011569039 0.016787893374491106 0 +163 0 -2.2131763 0.018867977 0.027480813524908082 0 +165 0 -1.9289668 0.026847113 0.039261617630834335 0 +166 1 5.3712273 0.9965933 0.0049232249583664065 1 +168 0 -2.6042523 0.011569039 0.016787893374491106 0 +170 0 -2.8614266 0.008373139 0.012130744104216924 0 +172 0 -3.1699133 0.005674961 0.0082105575200780885 0 +175 1 5.180047 0.99566144 0.0062728400019319355 1 +178 0 -2.400572 0.014932421 0.021705393169323427 0 +182 0 -1.959885 0.02584031 0.037769808679167967 0 +184 1 3.955361 0.9797846 0.029463467380837841 1 +185 0 -2.8716645 0.008265896 0.011974726218235726 0 +186 1 3.985546 0.98052996 0.028366376544727477 1 +190 1 7.4657273 0.9997608 0.00034512446635598279 1 +193 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +194 0 -2.6042523 0.011569039 0.016787893374491106 0 +195 0 -2.400572 0.014932421 0.021705393169323427 0 +197 0 -1.8385305 0.030015768 0.043966800471659635 0 +200 1 6.0277863 0.99851704 0.0021410521205442053 1 +203 0 -2.1858046 0.019522207 0.028443139723819993 0 +208 0 -3.0978432 0.006215379 0.0089948789544128712 0 +213 1 8.651912 0.99994695 7.6534309653049576E-05 1 +214 1 8.021141 0.9998818 0.00017053087656366316 1 +215 1 4.766493 0.9926871 0.010589041348733017 1 +217 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +220 0 -2.9111683 0.007864704 0.01139122322102276 0 +221 1 5.885932 0.9982249 0.0025631832715805789 1 +222 1 -1.0910791 0.074019454 3.7559516878709833 0 +224 1 5.8083696 0.9980415 0.0028282734477503191 1 +225 0 -3.1699133 0.005674961 0.0082105575200780885 0 +227 1 4.6679697 0.9917207 0.011994260047261323 1 +229 1 7.4762726 0.99976397 0.00034056584291574553 1 +230 1 3.553524 0.9667756 0.048747038591416381 1 +231 1 5.227359 0.9959133 0.0059079033199315846 1 +232 0 0.6768029 0.4299971 0.81095880149551824 1 +234 0 -1.2008077 0.065019004 0.096991053227062293 0 +235 0 ? ? ? 0 +236 1 6.9204206 0.9995221 0.00068964325415644641 1 +238 1 7.6631403 0.9998138 0.0002686619145078823 1 +243 0 -1.7283037 0.03436979 0.050457281588718518 0 +245 0 -1.7236625 0.034565903 0.050750314770831341 0 +251 1 4.627567 0.99128866 0.01262286539422033 1 +253 1 5.30029 0.9962734 0.0053863923631196743 1 +255 1 3.4353647 0.96160424 0.05648484121740667 1 +256 0 -2.8614266 0.008373139 0.012130744104216924 0 +261 1 6.7552137 0.9994106 0.00085053289008703693 1 +263 1 5.02655 0.99473274 0.007619137087775783 1 +264 1 3.0155754 0.93629414 0.094966268717484412 1 +265 0 -1.4002798 0.051216725 0.07584951677237968 0 +266 1 5.257723 0.9960673 0.00568489352489388 1 +270 1 4.0899086 0.982905 0.024876154420347706 1 +273 1 0.335526 0.3284555 1.606230137088891 1 +274 0 -2.414615 0.014672406 0.021324634644573313 0 +281 0 -2.697853 0.01028602 0.014916437796865623 0 +282 1 3.025592 0.93704855 0.093804289898579224 1 +286 1 8.930486 0.99996275 5.3745579490272704E-05 1 +287 0 -2.7086492 0.010147406 0.014714395385468908 0 +289 1 5.334999 0.9964335 0.0051545743258560924 1 +292 1 ? ? ? 0 +294 0 ? ? ? 0 +295 1 4.5509434 0.9904071 0.013906424460647474 1 +298 0 -1.5676326 0.041822422 0.061635041502073613 0 +302 1 8.872598 0.9999599 5.7873322786187624E-05 1 +305 1 5.7573247 0.9979107 0.003017407196179228 1 +306 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +307 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +310 0 -2.922761 0.0077506816 0.011225428686079548 0 +313 0 -3.299302 0.004819355 0.0069696676947273014 0 +315 0 ? ? ? 0 +318 0 -2.8691468 0.008292142 0.012012908069216954 0 +320 1 3.933642 0.9792311 0.030278688471899282 1 +322 0 -2.6042523 0.011569039 0.016787893374491106 0 +324 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +325 0 -2.0716977 0.022497272 0.032827366385524651 0 +326 1 2.324747 0.8594242 0.21855773716465846 1 +330 1 2.9209404 0.92874014 0.10665309970829556 1 +334 1 4.1128087 0.9833867 0.024169258370237343 1 +335 0 -3.299302 0.004819355 0.0069696676947273014 0 +337 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +339 1 3.6997185 0.972248 0.040603706580857601 1 +340 1 3.8200495 0.97608596 0.034919887823900035 1 +341 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +342 0 -3.1115277 0.006108973 0.0088404158141125212 0 +345 0 -3.299302 0.004819355 0.0069696676947273014 0 +351 0 -2.8278565 0.008734563 0.012656667304014846 0 +356 1 -0.28485298 0.18199818 2.4580040844636111 0 +357 1 6.906702 0.9995137 0.00070177387936437943 1 +359 1 3.2611713 0.9525508 0.070132017498008001 1 +362 0 -1.708355 0.035220396 0.051728686397833394 0 +363 0 -0.7258245 0.11276888 0.1726181270358976 0 +364 0 -2.8278565 0.008734563 0.012656667304014846 0 +365 0 -2.9772017 0.0072367755 0.010478420978548253 0 +367 1 6.7865734 0.9994336 0.0008174070864974929 1 +369 0 -2.8500295 0.00849415 0.012306810184418862 0 +372 0 -2.3130798 0.016657654 0.024234323683505889 0 +374 0 -2.5725198 0.012038953 0.017473933576498117 0 +375 0 -3.299302 0.004819355 0.0069696676947273014 0 +380 0 -3.299302 0.004819355 0.0069696676947273014 0 +382 0 -1.9081124 0.027547635 0.04030051281996655 0 +385 0 -1.8123745 0.030997913 0.045428322514214421 0 +386 1 3.5814629 0.96789634 0.047075545827350976 1 +390 0 -3.0397086 0.006688326 0.0096816280974065787 0 +393 0 -3.3468409 0.0045383475 0.0065623538118057389 0 +394 0 -2.53023 0.012694621 0.018431708315918838 0 +397 0 -2.6262507 0.0112539595 0.016328082658637364 0 +400 1 5.6133165 0.99749255 0.0036220259996406776 1 +401 0 -2.8614266 0.008373139 0.012130744104216924 0 +402 0 -1.6083863 0.03979731 0.058589116865399464 0 +403 0 -2.0118616 0.02423004 0.035387026571122526 0 +405 0 -3.1699133 0.005674961 0.0082105575200780885 0 +407 0 -3.1699133 0.005674961 0.0082105575200780885 0 +408 0 -1.7974956 0.031570416 0.046280942730234045 0 +410 0 -3.1699133 0.005674961 0.0082105575200780885 0 +411 0 ? ? ? 0 +412 1 5.1801887 0.9956622 0.0062717172439807865 1 +417 0 -3.1699133 0.005674961 0.0082105575200780885 0 +420 0 -1.3703549 0.053094853 0.078708178587860597 0 +421 1 6.7476797 0.99940497 0.00085870690667158795 1 +424 0 -2.8614266 0.008373139 0.012130744104216924 0 +425 1 8.440741 0.9999306 0.00010009737521057061 1 +426 0 -1.1437707 0.06956263 0.10401905088360545 0 +427 1 3.029043 0.9373065 0.093407171483572096 1 +431 0 -1.600387 0.04018725 0.059175118411096385 0 +432 0 -2.0669918 0.022629047 0.033021865599754556 0 +433 0 -2.1939526 0.019325165 0.028153237555685478 0 +435 1 4.930357 0.9940526 0.00860591734976527 1 +437 0 -2.6262507 0.0112539595 0.016328082658637364 0 +438 0 -1.9525015 0.026077349 0.038120896576848001 0 +443 0 -3.0108025 0.0069366256 0.01004230558528737 0 +444 0 -1.733223 0.034163095 0.05014850447576482 0 +445 0 -3.1115277 0.006108973 0.0088404158141125212 0 +446 0 -3.299302 0.004819355 0.0069696676947273014 0 +447 0 -2.3910747 0.01511084 0.021966723054784913 0 +448 0 -3.3468409 0.0045383475 0.0065623538118057389 0 +458 0 -2.212277 0.018889125 0.027511911404985143 0 +459 0 -2.033479 0.023589509 0.034440298131755802 0 +460 0 -2.1183214 0.021231595 0.030960564068005136 0 +461 0 -1.7272936 0.034412377 0.050520910663531066 0 +462 0 -1.9270117 0.026912041 0.03935787645297139 0 +463 0 -2.435658 0.014291105 0.02076644985353357 0 +468 0 -2.6262507 0.0112539595 0.016328082658637364 0 +469 0 -2.9698434 0.0073042074 0.010576417192379716 0 +470 0 -2.7434447 0.009713116 0.014081563613944885 0 +471 0 -1.9270117 0.026912041 0.03935787645297139 0 +472 0 -2.1959991 0.01927598 0.028080882086838562 0 +473 0 -2.6262507 0.0112539595 0.016328082658637364 0 +475 0 -2.8614266 0.008373139 0.012130744104216924 0 +476 0 -2.4474528 0.014081671 0.020459952428535714 0 +477 0 -2.6262507 0.0112539595 0.016328082658637364 0 +478 0 -2.2293873 0.018490694 0.026926149164253149 0 +479 1 5.2473927 0.99601555 0.0057598306401665739 1 +481 0 -1.4591032 0.04770667 0.07052206859546864 0 +485 0 -2.51326 0.012967533 0.018830554697715829 0 +486 0 -2.7434447 0.009713116 0.014081563613944885 0 +488 1 1.2017398 0.59499276 0.74905598903732029 1 +490 0 -3.299302 0.004819355 0.0069696676947273014 0 +491 1 4.4116144 0.98857194 0.01658213512685398 1 +494 0 0.19906044 0.29143023 0.49701818839808581 1 +496 0 -3.3468409 0.0045383475 0.0065623538118057389 0 +498 0 -2.4918776 0.013319657 0.019345327864414775 0 +499 0 -2.4918776 0.013319657 0.019345327864414775 0 +500 0 -1.959885 0.02584031 0.037769808679167967 0 +503 0 -2.400572 0.014932421 0.021705393169323427 0 +505 0 -2.5028172 0.013138341 0.019080236704647379 0 +506 1 6.2301693 0.99885267 0.0016561973072822019 1 +508 0 -2.3910747 0.01511084 0.021966723054784913 0 +509 0 -3.1115277 0.006108973 0.0088404158141125212 0 +511 0 -2.280314 0.017352993 0.025254839289623091 0 +512 0 -2.3910747 0.01511084 0.021966723054784913 0 +515 1 5.601015 0.99745315 0.0036790102734250547 1 +516 0 -3.3468409 0.0045383475 0.0065623538118057389 0 +518 0 -2.5321863 0.012663528 0.018386274342810423 0 +524 0 -2.7270536 0.009915355 0.014376224049015385 0 +525 0 -2.688909 0.010402273 0.015085908849059596 0 +526 0 -2.6262507 0.0112539595 0.016328082658637364 0 +530 1 4.0611787 0.9822811 0.025792170722406128 1 +536 0 -2.1858046 0.019522207 0.028443139723819993 0 +537 0 -1.9876449 0.02496774 0.036478141155914565 0 +542 0 -2.0680609 0.022599043 0.032977578199036302 0 +543 0 -2.4918776 0.013319657 0.019345327864414775 0 +545 0 -2.280314 0.017352993 0.025254839289623091 0 +550 0 -2.7270536 0.009915355 0.014376224049015385 0 +551 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +552 0 -1.7737393 0.032505758 0.047675019900819467 0 +553 0 -0.4828031 0.14751783 0.23025843519524158 0 +554 0 -2.8614266 0.008373139 0.012130744104216924 0 +555 0 -0.72908664 0.112355135 0.17194550674939729 0 +556 0 -1.8535137 0.029466813 0.043150549355838286 0 +562 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +564 0 -2.1686568 0.019943342 0.02906293911727361 0 +567 0 -1.9768028 0.025305066 0.036977349302418229 0 +568 1 3.127718 0.9442777 0.08271688898461392 1 +570 1 4.791981 0.9929183 0.010253062861046026 1 +571 1 6.3373704 0.9989985 0.001445550213597871 1 +572 0 -2.7270536 0.009915355 0.014376224049015385 0 +573 0 -3.1699133 0.005674961 0.0082105575200780885 0 +574 1 3.537343 0.96610934 0.049741625551089678 1 +575 0 -1.9876449 0.02496774 0.036478141155914565 0 +576 0 -2.280314 0.017352993 0.025254839289623091 0 +579 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +580 0 -2.0864134 0.022090027 0.032226439159456409 0 +583 0 -2.8614266 0.008373139 0.012130744104216924 0 +585 0 -3.299302 0.004819355 0.0069696676947273014 0 +587 0 -2.0669918 0.022629047 0.033021865599754556 0 +588 1 3.0994945 0.94236183 0.08564698778891873 1 +589 0 -2.3910747 0.01511084 0.021966723054784913 0 +591 1 3.9345682 0.9792551 0.030243387214391203 1 +592 1 3.106784 0.9428625 0.084880684135415621 1 +595 0 -2.280314 0.017352993 0.025254839289623091 0 +596 0 -2.3130798 0.016657654 0.024234323683505889 0 +597 0 -1.7976105 0.031565957 0.046274299798906078 0 +598 0 -2.7270536 0.009915355 0.014376224049015385 0 +599 0 -1.7464848 0.033611834 0.049325306851713629 0 +601 0 -3.2164953 0.005350794 0.007740291094224745 0 +603 1 2.4305255 0.87488085 0.19284154464442543 1 +605 1 5.5752525 0.9973687 0.0038011762787992279 1 +608 1 6.1609693 0.99874747 0.0018081543347426264 1 +610 1 4.256276 0.9861149 0.020172310679415974 1 +611 1 4.1918106 0.98494846 0.021879866566343848 1 +615 0 -2.1282535 0.020971112 0.030576665061562713 0 +616 0 -2.7270536 0.009915355 0.014376224049015385 0 +620 0 -2.7270536 0.009915355 0.014376224049015385 0 +623 0 -3.299302 0.004819355 0.0069696676947273014 0 +625 0 -1.6549138 0.0376002 0.055291753119665971 0 +626 1 3.2866921 0.9539941 0.067947758081130266 1 +628 0 -3.1115277 0.006108973 0.0088404158141125212 0 +630 0 -1.7293918 0.034323964 0.050388818903414519 0 +631 0 -2.280314 0.017352993 0.025254839289623091 0 +632 0 -3.299302 0.004819355 0.0069696676947273014 0 +635 0 -2.3237798 0.016436573 0.023910004561854832 0 +636 1 5.4205694 0.9967995 0.0046247087970721114 1 +637 0 -1.247425 0.0615121 0.091589952710676173 0 +640 0 -2.3239968 0.016432121 0.023903474771901047 0 +643 0 -3.299302 0.004819355 0.0069696676947273014 0 +646 0 -2.869265 0.008290907 0.012011111540480535 0 +647 0 -3.05244 0.006581785 0.0095268953013750626 0 +648 1 6.2857337 0.99893075 0.0015434236851713337 1 +650 0 -1.8218272 0.03063943 0.044894695439925601 0 +651 0 -2.7272372 0.009913066 0.014372889731158276 0 +655 0 -2.7270536 0.009915355 0.014376224049015385 0 +658 1 5.1765275 0.99564207 0.006300909234690056 1 +659 0 -3.299302 0.004819355 0.0069696676947273014 0 +662 0 -2.790698 0.009152671 0.013265313083523538 0 +663 0 -2.790698 0.009152671 0.013265313083523538 0 +664 0 -2.5114403 0.012997137 0.018873825932234793 0 +666 0 -1.9454997 0.02630409 0.038456813549173768 0 +667 0 -2.6042523 0.011569039 0.016787893374491106 0 +669 1 5.3159976 0.9963468 0.0052801449977246768 1 +671 0 -2.322586 0.016461095 0.023945973578975963 0 +672 0 -2.8278565 0.008734563 0.012656667304014846 0 +673 0 -1.8290145 0.03036955 0.0444930893633559 0 +674 0 -3.1699133 0.005674961 0.0082105575200780885 0 +675 0 -2.0441408 0.023279706 0.033982622004851887 0 +676 0 -2.9698434 0.0073042074 0.010576417192379716 0 +677 0 -2.3910747 0.01511084 0.021966723054784913 0 +684 0 -3.299302 0.004819355 0.0069696676947273014 0 +686 0 -3.299302 0.004819355 0.0069696676947273014 0 +687 0 -2.4359035 0.014286714 0.020760022876868865 0 +690 0 -3.05244 0.006581785 0.0095268953013750626 0 +695 0 -3.1115277 0.006108973 0.0088404158141125212 0 diff --git a/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer-out.txt b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer-out.txt new file mode 100644 index 0000000000..242ef9cd77 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer-out.txt @@ -0,0 +1,118 @@ +maml.exe TrainTest test=%Data% tr=WeightedEnsemble{bp=svm bp=ap nm=20 tp=-} dout=%Output% loader=Text{col=Label:BL:0 col=Features:R4:1-9} data=%Data% out=%Output% seed=1 +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Training 20 learners for the batch 1 +Beginning training model 1 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 26 instances with missing features during training (over 1 iterations; 26 inst/iter) +Trainer 1 of 20 finished in %Time% +Beginning training model 2 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 150 instances with missing features during training (over 10 iterations; 15 inst/iter) +Trainer 2 of 20 finished in %Time% +Beginning training model 3 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 11 instances with missing features during training (over 1 iterations; 11 inst/iter) +Trainer 3 of 20 finished in %Time% +Beginning training model 4 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 120 instances with missing features during training (over 10 iterations; 12 inst/iter) +Trainer 4 of 20 finished in %Time% +Beginning training model 5 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 18 instances with missing features during training (over 1 iterations; 18 inst/iter) +Trainer 5 of 20 finished in %Time% +Beginning training model 6 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 150 instances with missing features during training (over 10 iterations; 15 inst/iter) +Trainer 6 of 20 finished in %Time% +Beginning training model 7 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 24 instances with missing features during training (over 1 iterations; 24 inst/iter) +Trainer 7 of 20 finished in %Time% +Beginning training model 8 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 180 instances with missing features during training (over 10 iterations; 18 inst/iter) +Trainer 8 of 20 finished in %Time% +Beginning training model 9 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 15 instances with missing features during training (over 1 iterations; 15 inst/iter) +Trainer 9 of 20 finished in %Time% +Beginning training model 10 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 180 instances with missing features during training (over 10 iterations; 18 inst/iter) +Trainer 10 of 20 finished in %Time% +Beginning training model 11 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 14 instances with missing features during training (over 1 iterations; 14 inst/iter) +Trainer 11 of 20 finished in %Time% +Beginning training model 12 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 130 instances with missing features during training (over 10 iterations; 13 inst/iter) +Trainer 12 of 20 finished in %Time% +Beginning training model 13 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 14 instances with missing features during training (over 1 iterations; 14 inst/iter) +Trainer 13 of 20 finished in %Time% +Beginning training model 14 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 170 instances with missing features during training (over 10 iterations; 17 inst/iter) +Trainer 14 of 20 finished in %Time% +Beginning training model 15 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 24 instances with missing features during training (over 1 iterations; 24 inst/iter) +Trainer 15 of 20 finished in %Time% +Beginning training model 16 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 120 instances with missing features during training (over 10 iterations; 12 inst/iter) +Trainer 16 of 20 finished in %Time% +Beginning training model 17 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 12 instances with missing features during training (over 1 iterations; 12 inst/iter) +Trainer 17 of 20 finished in %Time% +Beginning training model 18 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 190 instances with missing features during training (over 10 iterations; 19 inst/iter) +Trainer 18 of 20 finished in %Time% +Beginning training model 19 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 23 instances with missing features during training (over 1 iterations; 23 inst/iter) +Trainer 19 of 20 finished in %Time% +Beginning training model 20 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 180 instances with missing features during training (over 10 iterations; 18 inst/iter) +Trainer 20 of 20 finished in %Time% +Training calibrator. +Warning: The predictor produced non-finite prediction values on 16 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3499 (239.0/(239.0+444.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 231 | 8 | 0.9665 + negative || 9 | 435 | 0.9797 + ||====================== +Precision || 0.9625 | 0.9819 | +OVERALL 0/1 ACCURACY: 0.975110 +LOG LOSS/instance: 0.112669 +Test-set entropy (prior Log-Loss/instance): 0.934003 +LOG-LOSS REDUCTION (RIG): 0.879370 +AUC: 0.996287 + +OVERALL RESULTS +--------------------------------------- +AUC: 0.996287 (0.0000) +Accuracy: 0.975110 (0.0000) +Positive precision: 0.962500 (0.0000) +Positive recall: 0.966527 (0.0000) +Negative precision: 0.981941 (0.0000) +Negative recall: 0.979730 (0.0000) +Log-loss: 0.112669 (0.0000) +Log-loss reduction: 0.879370 (0.0000) +F1 Score: 0.964509 (0.0000) +AUPRC: 0.992508 (0.0000) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer-rp.txt b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer-rp.txt new file mode 100644 index 0000000000..3895e807ec --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer-rp.txt @@ -0,0 +1,4 @@ +WeightedEnsemble +AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC /bp /nm Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.996287 0.97511 0.9625 0.966527 0.981941 0.97973 0.112669 0.87937 0.964509 0.992508 svm,ap 20 WeightedEnsemble %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsemble{bp=svm bp=ap nm=20 tp=-} dout=%Output% loader=Text{col=Label:BL:0 col=Features:R4:1-9} data=%Data% out=%Output% seed=1 /bp:svm,ap;/nm:20 + diff --git a/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer.txt b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer.txt new file mode 100644 index 0000000000..e02bc0a582 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/netcoreapp/WE-Hetero-TrainTest-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +0 0 -3.2653847 0.01606247 0.023361373063565786 0 +1 0 2.3942997 0.89365447 3.233168642152529 1 +2 0 -3.5103545 0.012305525 0.017863255262247732 0 +3 0 2.330025 0.8867252 3.1421009895524099 1 +4 0 -2.964328 0.022249075 0.032461098556400524 0 +5 1 9.490736 0.9999526 6.836472348564471E-05 1 +6 0 -1.6140783 0.09167324 0.13871671132075303 0 +7 0 -3.8577213 0.008421281 0.012200786382637996 0 +8 0 -3.6269126 0.01083681 0.015719541380870845 0 +9 0 -3.6241956 0.010868987 0.015766472600222039 0 +10 0 -4.5891733 0.0037754006 0.0054570594742051276 0 +11 0 -4.3108807 0.005125152 0.0074130440548942798 0 +12 1 -0.49860764 0.25677502 1.9614232274515342 0 +13 0 -3.926801 0.0078081824 0.011309035197640359 0 +14 1 6.8671675 0.99914455 0.0012346756770163963 1 +15 1 1.0072589 0.64530325 0.63195080484223587 1 +16 0 -3.7142038 0.009851713 0.014283491630889444 0 +17 0 -3.5423806 0.011883428 0.017246843094017798 0 +18 1 6.1876116 0.99819136 0.0026116832932766227 1 +19 0 -2.787673 0.026907591 0.039351279130343614 0 +20 1 4.5410404 0.98897916 0.015987973243636426 1 +21 1 6.0276127 0.99784297 0.0031153011881840933 1 +22 0 -3.9912 0.007276599 0.010536293898534796 0 +23 1 ? ? ? 0 +24 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +25 1 0.9573393 0.63259983 0.66063492529981271 1 +26 0 -3.9880347 0.007301866 0.010573014495608041 0 +27 0 -3.437208 0.013325897 0.019354451635261342 0 +28 0 -4.3108807 0.005125152 0.0074130440548942798 0 +29 0 -4.175107 0.005948365 0.0086073021192332854 0 +30 0 -4.12895 0.006257154 0.0090555256618609532 0 +31 0 -4.180897 0.005910713 0.0085526577863213431 0 +32 1 5.927128 0.9975907 0.0034800492560580422 1 +33 0 -4.0466037 0.006848121 0.0099137343988522971 0 +34 0 -3.6587863 0.010466299 0.015179252284484622 0 +35 0 -4.3108807 0.005125152 0.0074130440548942798 0 +36 1 7.264369 0.9994479 0.0007967576197584538 1 +37 0 -1.5557506 0.09717395 0.14748005096538608 0 +38 1 4.131503 0.9827931 0.025040376709636651 1 +39 1 1.3442421 0.725162 0.46362470982637882 1 +40 0 ? ? ? 0 +41 1 2.4566755 0.90001917 0.1519723662053202 1 +42 1 6.9449177 0.9992148 0.0011332086699996601 1 +43 1 0.5260675 0.51689655 0.95205253324632289 1 +44 1 6.58985 0.9988388 0.0016762564398381671 1 +45 0 -4.5121536 0.004108833 0.0059400048264421166 0 +46 1 3.6237736 0.97025704 0.043561093211102381 1 +47 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +48 0 -2.964328 0.022249075 0.032461098556400524 0 +49 1 4.9304733 0.9928004 0.010424376994896773 1 +50 1 1.8948932 0.8288765 0.27077094226606374 1 +51 1 -0.27271032 0.30712578 1.7030984923805075 0 +52 1 4.192661 0.9838977 0.023419795144202202 1 +53 1 5.4079657 0.995736 0.0061648001441107422 1 +54 1 4.8227606 0.9918994 0.011734329225674151 1 +55 1 4.216007 0.9843007 0.02282901387797074 1 +56 1 5.273701 0.9950586 0.007146610574688802 1 +57 1 1.0327673 0.6517176 0.61768113151824056 1 +58 1 1.3259256 0.72111654 0.47169565628725918 1 +59 1 1.067169 0.66028154 0.59884678517541101 1 +60 1 1.9315443 0.83453554 0.2609546060375782 1 +61 0 -4.3741302 0.0047813808 0.0069146182292416161 0 +62 1 5.711467 0.9969455 0.0044134553486145471 1 +63 1 0.13153958 0.40911457 1.289423179358389 1 +64 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +65 1 2.4733298 0.9016603 0.14934405654224964 1 +66 0 -3.5423806 0.011883428 0.017246843094017798 0 +67 1 3.0271854 0.94410795 0.082976267459593273 1 +68 1 7.6869254 0.9996535 0.00049995419224947387 1 +69 0 -4.360554 0.00485318 0.0070187042540280362 0 +70 0 -2.8683703 0.024672236 0.036040969757435599 0 +71 1 5.8321104 0.9973251 0.0038642031538066459 1 +72 0 -2.5583074 0.03438858 0.05048535551266551 0 +73 1 6.377676 0.99853295 0.0021180585209483065 1 +74 1 2.3969789 0.893935 0.16175812130252948 1 +75 0 -3.6002493 0.011156697 0.016186172128346488 0 +76 0 -3.7219474 0.00976873 0.014162586791980635 0 +77 0 -2.843206 0.025349153 0.037042606089538327 0 +78 0 -3.2013326 0.01721828 0.025057072303905244 0 +79 0 -4.2326174 0.005584737 0.0080796545461867403 0 +80 0 -3.0660248 0.019934934 0.029050562180483869 0 +81 0 -3.5765634 0.011448682 0.016612233748243246 0 +82 0 -3.1724813 0.017765233 0.0258602065143585 0 +83 0 -2.6362126 0.031646147 0.046393766241014454 0 +84 1 6.649848 0.99891305 0.0015689906725689917 1 +85 1 5.1913624 0.99459124 0.007824376063047667 1 +86 1 1.3384473 0.7238861 0.46616543539216543 1 +87 1 5.110711 0.9940909 0.0085502951855127522 1 +88 0 -3.5423806 0.011883428 0.017246843094017798 0 +89 0 -4.098324 0.0064707827 0.0093657005784377242 0 +90 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +91 0 -4.0253825 0.0070091966 0.01014773855782411 0 +92 0 -3.5423806 0.011883428 0.017246843094017798 0 +93 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +94 0 -4.180897 0.005910713 0.0085526577863213431 0 +95 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +96 0 -4.304512 0.0051611005 0.0074651750173645066 0 +97 0 -3.2653847 0.01606247 0.023361373063565786 0 +98 1 7.1164656 0.99935013 0.00093786820013878381 1 +99 1 7.562501 0.99960256 0.00057350413335023013 1 +100 1 4.2504015 0.98487633 0.021985509976149925 1 +101 1 -0.70509744 0.21575198 2.2125543237686305 0 +102 0 -3.2099671 0.01705783 0.024821555415505958 0 +103 1 0.6878574 0.5612156 0.8333730357345136 1 +104 1 9.079937 0.99992543 0.0001075791590066161 1 +105 1 1.5973396 0.77720076 0.36364078606152644 1 +106 1 7.660897 0.99964345 0.00051449183649684158 1 +107 1 4.957925 0.9930137 0.01011450210336301 1 +108 0 -4.3402457 0.004962593 0.0071773319512800447 0 +109 1 4.805126 0.99174154 0.011963912139766504 1 +110 0 -2.7174697 0.029011479 0.042473854120173266 0 +111 1 3.0339077 0.944498 0.082380349352960389 1 +112 1 5.526126 0.99625516 0.0054128043766188405 1 +113 1 7.2231 0.9994222 0.00083384083165018218 1 +114 0 -2.6591587 0.030879542 0.045252096920480538 0 +115 0 -3.6688507 0.010351932 0.015012519544694238 0 +116 0 -0.6106701 0.233899 0.38439348540221963 0 +117 1 6.245225 0.9983025 0.0024510278321958367 1 +118 0 -4.0302415 0.0069719865 0.010093677869498451 0 +119 0 -3.4192936 0.01358826 0.019738124381236736 0 +120 0 -4.0978036 0.0064744735 0.0093710600121635016 0 +121 0 -3.1186867 0.018830981 0.027426414388720999 0 +122 1 7.959108 0.9997434 0.00037024015945535141 1 +123 1 2.2560513 0.87826604 0.18727008061830744 1 +124 1 5.382405 0.9956146 0.0063407252985224448 1 +125 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +126 1 5.499481 0.9961439 0.0055739626604518182 1 +127 0 -3.8193765 0.008781962 0.012725653416849284 0 +128 1 4.0050287 0.9802676 0.028752478431827138 1 +129 0 -4.1953506 0.0058177593 0.0084177625850741979 0 +130 0 -2.8683703 0.024672236 0.036040969757435599 0 +131 0 -4.180897 0.005910713 0.0085526577863213431 0 +132 1 6.6597075 0.9989248 0.0015520320478416001 1 +133 0 -3.8269148 0.008709866 0.012620722473631272 0 +134 0 -3.8534496 0.008460724 0.012258174209428919 0 +135 0 -2.6051593 0.032712977 0.047984051442468541 0 +136 0 -3.7142038 0.009851713 0.014283491630889444 0 +137 0 -4.163023 0.006027712 0.0087224649837247988 0 +138 0 -3.3817904 0.014154118 0.020565967564274219 0 +139 0 ? ? ? 0 +140 0 -4.163023 0.006027712 0.0087224649837247988 0 +141 0 -4.4408646 0.0044435263 0.0064249385978414108 0 +142 1 3.6463163 0.97096634 0.042506812877205175 1 +143 0 -3.6688507 0.010351932 0.015012519544694238 0 +144 0 -4.3108807 0.005125152 0.0074130440548942798 0 +145 0 ? ? ? 0 +146 1 0.7448549 0.57663524 0.79426908393757356 1 +147 0 -4.2958665 0.0052103032 0.0075365295361707703 0 +148 0 -1.5269843 0.099993885 0.15199329046725604 0 +149 1 9.159508 0.99993175 9.8463427589617192E-05 1 +150 0 -4.5891733 0.0037754006 0.0054570594742051276 0 +151 1 3.6840632 0.97211754 0.040797327590637526 1 +152 1 7.6329136 0.9996323 0.00053057803961755552 1 +153 0 -3.2679832 0.016017227 0.023295036240513722 0 +154 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +155 1 2.7025309 0.92191553 0.11729352249888303 1 +156 0 -4.134472 0.0062193903 0.0090007021066693289 0 +157 0 -4.180897 0.005910713 0.0085526577863213431 0 +158 0 ? ? ? 0 +159 1 10.04513 0.9999743 3.7062737326457185E-05 1 +160 1 7.503611 0.9995759 0.00061195805135972042 1 +161 0 -3.3780956 0.014211106 0.0206493675057071 0 +162 0 -3.8193765 0.008781962 0.012725653416849284 0 +163 0 -3.1213367 0.01877704 0.027347103564814897 0 +164 0 ? ? ? 0 +165 0 -3.0690737 0.019869326 0.028953987881998787 0 +166 1 6.444971 0.9986378 0.0019665855015772657 1 +167 1 6.233284 0.99828005 0.002483502051107291 1 +168 0 -3.8193765 0.008781962 0.012725653416849284 0 +169 0 -4.6255016 0.0036276237 0.0052430703738574847 0 +170 0 -4.163023 0.006027712 0.0087224649837247988 0 +171 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +172 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +173 1 10.937983 0.9999904 1.3844669823695805E-05 1 +174 1 4.1922584 0.98389065 0.023430108221276183 1 +175 1 5.9543967 0.99766195 0.0033770451247637162 1 +176 0 -4.180897 0.005910713 0.0085526577863213431 0 +177 1 4.079586 0.9817973 0.026502928108117836 1 +178 0 -3.5423806 0.011883428 0.017246843094017798 0 +179 1 1.358726 0.7283351 0.45732575615903537 1 +180 0 -4.5891733 0.0037754006 0.0054570594742051276 0 +181 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +182 0 -2.787673 0.026907591 0.039351279130343614 0 +183 1 6.992011 0.9992545 0.0010758945834295176 1 +184 1 4.773998 0.9914555 0.01238008094867033 1 +185 0 -4.259161 0.0054244464 0.0078471244057280662 0 +186 1 4.1567426 0.9832577 0.024358499727650511 1 +187 1 9.617186 0.9999588 5.9421229566932906E-05 1 +188 1 6.1444263 0.99810326 0.0027390143775226426 1 +189 0 -3.7622724 0.009347656 0.013549242633369745 0 +190 1 9.311988 0.9999423 8.3242004459984032E-05 1 +191 1 8.432387 0.99984777 0.00021963856339410703 1 +192 0 -3.437208 0.013325897 0.019354451635261342 0 +193 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +194 0 -3.8193765 0.008781962 0.012725653416849284 0 +195 0 -3.5423806 0.011883428 0.017246843094017798 0 +196 0 4.806697 0.9917557 6.9223914725779769 1 +197 0 -2.4172082 0.03994919 0.058817333240511131 0 +198 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +199 0 -3.9912 0.007276599 0.010536293898534796 0 +200 1 8.171349 0.9997969 0.00029300219717303392 1 +201 1 7.0345078 0.9992887 0.0010265856373871611 1 +202 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +203 0 -3.2653847 0.01606247 0.023361373063565786 0 +204 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +205 1 9.044386 0.99992245 0.00011187905230883949 1 +206 1 5.3570633 0.9954908 0.0065201271374746485 1 +207 0 -4.5891733 0.0037754006 0.0054570594742051276 0 +208 0 -4.5891733 0.0037754006 0.0054570594742051276 0 +209 0 -3.3103278 0.015297387 0.022240008119875913 0 +210 1 10.73283 0.99998796 1.7370352308075089E-05 1 +211 1 7.4396753 0.9995449 0.0006566932054570294 1 +212 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +213 1 11.035078 0.99999136 1.2468796069497877E-05 1 +214 1 10.208881 0.99997854 3.0957209291316721E-05 1 +215 1 6.245116 0.99830234 0.0024512862448457153 1 +216 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +217 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +218 1 6.0978594 0.9980035 0.002883244619054837 1 +219 0 -2.2632232 0.04699839 0.069449442047435667 0 +220 0 -4.3542194 0.0048870486 0.0070678052650200484 0 +221 1 7.3951244 0.99952203 0.00068972928660021061 1 +222 1 -2.0358088 0.059601545 4.0685064527685437 0 +223 1 4.0871634 0.9819461 0.026284243384590233 1 +224 1 7.2424936 0.9994344 0.00081620252615342435 1 +225 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +226 1 7.48438 0.99956685 0.00062503433754394636 1 +227 1 6.233205 0.9982799 0.0024836743300799221 1 +228 0 -4.5891733 0.0037754006 0.0054570594742051276 0 +229 1 10.090591 0.99997556 3.5256874230690643E-05 1 +230 1 4.2729473 0.98524237 0.021449427825201789 1 +231 1 6.6903114 0.99896044 0.0015005548038905881 1 +232 0 0.964937 0.6345457 1.4522370224952907 1 +233 1 4.6214604 0.98990536 0.014637495950801363 1 +234 0 -2.2420673 0.048054848 0.071049642177916109 0 +235 0 ? ? ? 0 +236 1 9.714638 0.999963 5.3401601415370602E-05 1 +237 1 5.7844906 0.9971813 0.0040722717432879136 1 +238 1 9.890707 0.99996954 4.3942236494581622E-05 1 +239 1 4.3297806 0.98612684 0.020154870358009745 1 +240 0 -1.6836672 0.08547793 0.12891011337435254 0 +241 0 -3.3089218 0.015320768 0.022274263986601432 0 +242 0 -4.180897 0.005910713 0.0085526577863213431 0 +243 0 -2.7772944 0.02720901 0.039798228263274164 0 +244 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +245 0 -2.551019 0.034656577 0.05088581904572332 0 +246 1 8.931614 0.9999122 0.00012667078313985091 1 +247 1 2.6447172 0.91719896 0.12469338209264234 1 +248 0 -2.7118673 0.029186089 0.042733312958056394 0 +249 0 ? ? ? 0 +250 0 -4.2502556 0.0054777074 0.007924384853928686 0 +251 1 5.724431 0.9969888 0.0043508357296072114 1 +252 0 3.280855 0.9571668 4.5451264226680612 1 +253 1 6.9449177 0.9992148 0.0011332086699996601 1 +254 1 5.711467 0.9969455 0.0044134553486145471 1 +255 1 3.3896317 0.9618255 0.056152934315026493 1 +256 0 -4.163023 0.006027712 0.0087224649837247988 0 +257 0 -3.9912 0.007276599 0.010536293898534796 0 +258 0 -3.8193765 0.008781962 0.012725653416849284 0 +259 0 2.5125067 0.9054266 3.402422052935989 1 +260 1 7.2856245 0.9994607 0.00077825939011553385 1 +261 1 9.371489 0.99994594 7.799624047875757E-05 1 +262 1 7.76371 0.99968165 0.00045935279012386564 1 +263 1 6.353957 0.99849415 0.0021741222097056657 1 +264 1 3.9517117 0.9790971 0.030476110431328859 1 +265 0 -2.3131797 0.04459069 0.065809158501979964 0 +266 1 6.232263 0.99827814 0.0024862585171381091 1 +267 1 1.6970851 0.7956729 0.32975264386834346 1 +268 1 7.200084 0.99940735 0.00085526520983337693 1 +269 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +270 1 5.017771 0.9934571 0.0094704558946413081 1 +271 0 -3.2653847 0.01606247 0.023361373063565786 0 +272 1 1.6970851 0.7956729 0.32975264386834346 1 +273 1 -0.08302784 0.35335246 1.5008201530548315 0 +274 0 -3.6550915 0.010508598 0.015240923338025363 0 +275 0 ? ? ? 0 +276 0 -3.9912 0.007276599 0.010536293898534796 0 +277 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +278 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +279 1 5.9754877 0.9977156 0.0032994736463515775 1 +280 0 -3.8193765 0.008781962 0.012725653416849284 0 +281 0 -4.0466037 0.006848121 0.0099137343988522971 0 +282 1 2.6876793 0.9207279 0.11915321749954444 1 +283 1 4.8468323 0.99210995 0.011428073251990358 1 +284 1 5.337451 0.9953926 0.0066624033876966004 1 +285 1 10.638187 0.99998665 1.926218548588174E-05 1 +286 1 12.19091 0.9999976 3.4396571176486151E-06 1 +287 0 -3.8534496 0.008460724 0.012258174209428919 0 +288 1 0.8549895 0.60598713 0.72264093713024813 1 +289 1 6.300803 0.9984033 0.0023053766008164661 1 +290 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +291 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +292 1 ? ? ? 0 +293 1 4.320032 0.9859789 0.020371319674727079 1 +294 0 ? ? ? 0 +295 1 5.3697605 0.99555326 0.0064296028627044708 1 +296 0 1.1073017 0.6701411 1.6000790676204995 1 +297 0 ? ? ? 0 +298 0 -1.8807185 0.06994541 0.10461269633998699 0 +299 1 4.882295 0.99241036 0.010991296384589865 1 +300 1 5.6863775 0.9968601 0.0045370637587865406 1 +301 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +302 1 12.254818 0.99999774 3.267674066997249E-06 1 +303 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +304 1 4.039583 0.9809916 0.0276873089888711 1 +305 1 7.158687 0.9993797 0.00089518939794833402 1 +306 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +307 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +308 1 5.4057903 0.99572575 0.0061796540653028253 1 +309 0 -2.1225212 0.05446038 0.080790183985325531 0 +310 0 -4.2326174 0.005584737 0.0080796545461867403 0 +311 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +312 1 2.7895415 0.9285517 0.10694589653145631 1 +313 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +314 0 -4.704957 0.0033241964 0.0048037905231660839 0 +315 0 ? ? ? 0 +316 1 2.5723104 0.9109269 0.13459284363516602 1 +317 1 6.9025745 0.99917734 0.0011873407314217728 1 +318 0 -4.326701 0.0050369254 0.0072851100604160307 0 +319 0 1.3230557 0.72047937 1.8389733255682628 1 +320 1 4.996194 0.99330056 0.0096977742317654117 1 +321 0 ? ? ? 0 +322 0 -3.8193765 0.008781962 0.012725653416849284 0 +323 1 3.6116142 0.96986747 0.044140477995098168 1 +324 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +325 0 -3.2402081 0.016507382 0.024013870445769369 0 +326 1 2.9760985 0.94105846 0.08764375146912004 1 +327 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +328 1 3.8758874 0.97731483 0.033104711543901995 1 +329 1 5.2012377 0.9946495 0.0077398216765183717 1 +330 1 4.254653 0.9849461 0.02188335878683351 1 +331 0 -2.941507 0.022803389 0.033279233976139994 0 +332 0 -2.5710278 0.033925634 0.049793846295632979 0 +333 1 3.4827366 0.965423 0.050766915308907944 1 +334 1 5.0232105 0.993496 0.0094139348463210797 1 +335 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +336 1 4.2282214 0.98450756 0.022525809242236003 1 +337 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +338 0 -4.704957 0.0033241964 0.0048037905231660839 0 +339 1 4.4128485 0.98732626 0.018401189309514714 1 +340 1 4.5462866 0.9890421 0.015896157405924871 1 +341 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +342 0 -4.4408646 0.0044435263 0.0064249385978414108 0 +343 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +344 1 6.7660546 0.9990437 0.0013803049211358605 1 +345 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +346 0 -2.5197663 0.03582873 0.052638650727444197 0 +347 0 -4.540222 0.003984052 0.0057592525983062234 0 +348 1 -0.012273073 0.37138504 1.4290124007333043 0 +349 1 2.2752178 0.8805086 0.18359099711192237 1 +350 0 -3.069314 0.019864162 0.02894638789646586 0 +351 0 -4.180897 0.005910713 0.0085526577863213431 0 +352 0 -0.07698059 0.35487828 0.63235669683728324 0 +353 1 6.50544 0.9987256 0.001839753075220707 1 +354 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +355 0 -3.4121685 0.013694022 0.019892816763642646 0 +356 1 -0.841964 0.19129993 2.3860917474901231 0 +357 1 9.989178 0.99997264 3.9470558303741813E-05 1 +358 1 5.0753384 0.9938572 0.0088895112699152994 1 +359 1 4.100359 0.98220235 0.025907818977461108 1 +360 1 11.307167 0.9999936 9.2011011621090467E-06 1 +361 1 4.063813 0.9814837 0.026963789668448576 1 +362 0 -2.7143378 0.029108964 0.042618705019448533 0 +363 0 -1.4885807 0.10387179 0.1582229440627041 0 +364 0 -4.180897 0.005910713 0.0085526577863213431 0 +365 0 -4.3108807 0.005125152 0.0074130440548942798 0 +366 1 10.271222 0.99998 2.8893374672045653E-05 1 +367 1 7.986121 0.9997509 0.00035940251222875554 1 +368 0 -4.175107 0.005948365 0.0086073021192332854 0 +369 0 -4.1233873 0.0062954286 0.0091110929417619214 0 +370 0 -3.0640237 0.01997811 0.029114120550850105 0 +371 0 -4.175107 0.005948365 0.0086073021192332854 0 +372 0 -3.3817904 0.014154118 0.020565967564274219 0 +373 0 -3.2544837 0.01625364 0.02364170371561946 0 +374 0 -3.6587863 0.010466299 0.015179252284484622 0 +375 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +376 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +377 0 -4.704957 0.0033241964 0.0048037905231660839 0 +378 0 -3.0130682 0.02110893 0.030779769130863978 0 +379 0 -1.9886942 0.06258234 0.093236115307430659 0 +380 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +381 1 7.0031204 0.99926364 0.0010627281554273494 1 +382 0 -3.0925539 0.0193711 0.028220814801589701 0 +383 0 -4.4408646 0.0044435263 0.0064249385978414108 0 +384 0 -4.4408646 0.0044435263 0.0064249385978414108 0 +385 0 -2.7200508 0.028931374 0.04235483911885983 0 +386 1 4.064584 0.98149914 0.026941097923102258 1 +387 0 -2.0610461 0.05806005 0.086293005977193429 0 +388 0 -4.120144 0.0063178525 0.0091436491391560324 0 +389 0 -2.8261507 0.025818188 0.037737046231875973 0 +390 0 -4.470003 0.004303564 0.0062221285147928303 0 +391 1 8.021375 0.9997604 0.00034572654977401063 1 +392 0 -3.9912 0.007276599 0.010536293898534796 0 +393 0 -4.82074 0.0029267597 0.0042286127291342535 0 +394 0 -3.878056 0.008236009 0.011931249562793627 0 +395 0 -3.9912 0.007276599 0.010536293898534796 0 +396 0 -3.8193765 0.008781962 0.012725653416849284 0 +397 0 -3.8860273 0.008164489 0.011827214924006168 0 +398 0 -3.7263813 0.009721528 0.0140938182217156 0 +399 0 -3.9938395 0.0072555942 0.010505768892170015 0 +400 1 7.209733 0.9994136 0.00084623079469699528 1 +401 0 -4.163023 0.006027712 0.0087224649837247988 0 +402 0 -2.4666376 0.03790946 0.055755424967277131 0 +403 0 -3.2359571 0.01658369 0.024125812989596599 0 +404 0 -3.9268205 0.007808015 0.011308792120877433 0 +405 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +406 0 -3.2917824 0.015608645 0.022696106958476066 0 +407 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +408 0 -2.9637718 0.022262428 0.032480801811266558 0 +409 0 -3.6587863 0.010466299 0.015179252284484622 0 +410 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +411 0 ? ? ? 0 +412 1 7.5729628 0.99960715 0.00056688018384502673 1 +413 0 -2.9279583 0.023138843 0.033774571618875787 0 +414 1 5.3245497 0.9953269 0.0067576939920419602 1 +415 0 -0.90735483 0.18038756 0.28698620721564849 0 +416 1 6.3837595 0.9985428 0.0021038491763182409 1 +417 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +418 0 -1.8329566 0.07345172 0.11006194189996231 0 +419 0 -3.8558304 0.0084387185 0.012226156748351875 0 +420 0 -2.203293 0.050049897 0.07407635876351483 0 +421 1 9.213475 0.9999357 9.278762391679814E-05 1 +422 0 -2.23039 0.048647597 0.071948247742412361 0 +423 0 -2.8683703 0.024672236 0.036040969757435599 0 +424 0 -4.163023 0.006027712 0.0087224649837247988 0 +425 1 11.08436 0.99999183 1.1780859684453884E-05 1 +426 0 -1.8131728 0.07495097 0.11239826175575356 0 +427 1 3.7117805 0.9729345 0.039585434891746686 1 +428 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +429 0 -4.3108807 0.005125152 0.0074130440548942798 0 +430 0 -3.9036994 0.008008121 0.011599784326242658 0 +431 0 -2.426865 0.039542593 0.058206458260191218 0 +432 0 -3.272923 0.015931563 0.023169444327594324 0 +433 0 -3.3588386 0.014511812 0.021089514904622398 0 +434 0 3.9872332 0.97988427 5.6355318697686085 1 +435 1 6.3551846 0.9984962 0.0021711940983063283 1 +436 1 4.0122623 0.98042136 0.028526172652705149 1 +437 0 -3.8860273 0.008164489 0.011827214924006168 0 +438 0 -2.9459198 0.022695165 0.033119465995260312 0 +439 0 -3.6090312 0.01105032 0.016030979963451743 0 +440 1 5.512389 0.9961982 0.0054953234556054968 1 +441 0 -1.4009573 0.11322032 0.1733523813753011 0 +442 0 -3.8031874 0.008938799 0.012953943516715532 0 +443 0 -4.5041137 0.004145287 0.0059928145324647268 0 +444 0 -2.2439585 0.047959503 0.070905151725369528 0 +445 0 -4.4408646 0.0044435263 0.0064249385978414108 0 +446 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +447 0 -3.6090312 0.01105032 0.016030979963451743 0 +448 0 -4.82074 0.0029267597 0.0042286127291342535 0 +449 1 7.769635 0.99968374 0.0004563421384379493 1 +450 0 -3.2185173 0.016900396 0.024590502376558559 0 +451 0 -3.6090312 0.01105032 0.016030979963451743 0 +452 0 -3.7536323 0.009436335 0.01367839324064603 0 +453 1 5.545363 0.9963335 0.0052993915030845867 1 +454 0 -3.9113288 0.007941535 0.011502949021920455 0 +455 1 0.9193044 0.6227945 0.68317187039427263 1 +456 1 7.4111066 0.9995304 0.00067768479439159895 1 +457 1 6.881011 0.99915755 0.0012159136400604084 1 +458 0 -3.2766178 0.015867786 0.023075947032752241 0 +459 0 -3.0443442 0.020407634 0.029746563705988448 0 +460 0 -3.069314 0.019864162 0.02894638789646586 0 +461 0 -2.9281356 0.023134422 0.033768041041859792 0 +462 0 -2.5544052 0.034531813 0.050699373131670437 0 +463 0 -3.7217422 0.009770921 0.014165779509094509 0 +464 0 -3.8860273 0.008164489 0.011827214924006168 0 +465 1 7.5764675 0.99960864 0.00056472955742893302 1 +466 1 7.0032444 0.9992637 0.0010626421007376144 1 +467 1 5.710013 0.9969406 0.004420528258786121 1 +468 0 -3.8860273 0.008164489 0.011827214924006168 0 +469 0 -4.3682995 0.0048120874 0.0069591320626229613 0 +470 0 -4.12895 0.006257154 0.0090555256618609532 0 +471 0 -2.5544052 0.034531813 0.050699373131670437 0 +472 0 -3.2288504 0.016712038 0.024314113995362596 0 +473 0 -3.8860273 0.008164489 0.011827214924006168 0 +474 0 -3.6090312 0.01105032 0.016030979963451743 0 +475 0 -4.163023 0.006027712 0.0087224649837247988 0 +476 0 -3.5536137 0.011738793 0.017035684405780797 0 +477 0 -3.8860273 0.008164489 0.011827214924006168 0 +478 0 -3.5190227 0.012189841 0.0176942887435638 0 +479 1 5.7542105 0.9970858 0.0042104258581785788 1 +480 0 -3.7467813 0.0095072435 0.013781670232746585 0 +481 0 -2.151782 0.052821834 0.078292269460707456 0 +482 1 11.169056 0.99999255 1.0748955721955734E-05 1 +483 1 8.170057 0.9997966 0.00029343224119461272 1 +484 0 -3.2766178 0.015867786 0.023075947032752241 0 +485 0 -3.8735213 0.008276972 0.011990839101656891 0 +486 0 -4.12895 0.006257154 0.0090555256618609532 0 +487 1 10.036007 0.999974 3.7492705063384392E-05 1 +488 1 0.83696985 0.6012308 0.73400917663757237 1 +489 1 -0.68922067 0.21873033 2.1927748078316514 0 +490 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +491 1 4.892073 0.9924912 0.010873805179547185 1 +492 0 -3.851954 0.008474577 0.012278331156502851 0 +493 1 7.652031 0.9996399 0.00051956714324732207 1 +494 0 -0.08959174 0.35169962 0.6252656774156361 0 +495 0 -4.12895 0.006257154 0.0090555256618609532 0 +496 0 -4.82074 0.0029267597 0.0042286127291342535 0 +497 0 -3.7960186 0.009009132 0.013056332112485152 0 +498 0 -3.7142038 0.009851713 0.014283491630889444 0 +499 0 -3.7142038 0.009851713 0.014283491630889444 0 +500 0 -2.787673 0.026907591 0.039351279130343614 0 +501 0 -3.7142038 0.009851713 0.014283491630889444 0 +502 0 -3.5765634 0.011448682 0.016612233748243246 0 +503 0 -3.5423806 0.011883428 0.017246843094017798 0 +504 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +505 0 -3.7836492 0.009131782 0.01323489774729354 0 +506 1 8.18573 0.99980015 0.0002883577299081774 1 +507 0 -3.9289443 0.007789885 0.011282430384316191 0 +508 0 -3.6090312 0.01105032 0.016030979963451743 0 +509 0 -4.4408646 0.0044435263 0.0064249385978414108 0 +510 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +511 0 -3.437208 0.013325897 0.019354451635261342 0 +512 0 -3.6090312 0.01105032 0.016030979963451743 0 +513 0 -4.12895 0.006257154 0.0090555256618609532 0 +514 1 7.3587427 0.9995025 0.00071794820497353992 1 +515 1 6.1920147 0.9982001 0.0025990197200290691 1 +516 0 -4.82074 0.0029267597 0.0042286127291342535 0 +517 0 -4.704957 0.0033241964 0.0048037905231660839 0 +518 0 -3.6929693 0.010082866 0.014620332791053558 0 +519 1 4.455138 0.9878969 0.017567581260147791 1 +520 0 -4.7431955 0.0031873218 0.0046056771836035793 0 +521 0 -3.7482767 0.009491722 0.01375906279519358 0 +522 1 3.5824206 0.9689119 0.045562623365110494 1 +523 1 5.427267 0.9958254 0.0060352666158911116 1 +524 0 -3.9912 0.007276599 0.010536293898534796 0 +525 0 -4.0253825 0.0070091966 0.01014773855782411 0 +526 0 -3.8860273 0.008164489 0.011827214924006168 0 +527 0 -3.5423806 0.011883428 0.017246843094017798 0 +528 0 -2.3386858 0.043407187 0.064023142466141603 0 +529 0 -3.851954 0.008474577 0.012278331156502851 0 +530 1 4.8951244 0.9925162 0.010837416039046229 1 +531 0 -3.2917824 0.015608645 0.022696106958476066 0 +532 0 -4.5891733 0.0037754006 0.0054570594742051276 0 +533 0 -3.9912 0.007276599 0.010536293898534796 0 +534 0 -4.3108807 0.005125152 0.0074130440548942798 0 +535 0 -3.780846 0.009159804 0.013275698921877643 0 +536 0 -3.2653847 0.01606247 0.023361373063565786 0 +537 0 -2.9279583 0.023138843 0.033774571618875787 0 +538 0 -3.7142038 0.009851713 0.014283491630889444 0 +539 0 -3.0149088 0.021067014 0.030717992997854295 0 +540 0 -3.128186 0.01863833 0.027143170511471044 0 +541 0 -4.163023 0.006027712 0.0087224649837247988 0 +542 0 -3.4282966 0.013455777 0.019544372621895852 0 +543 0 -3.7142038 0.009851713 0.014283491630889444 0 +544 0 -3.5302353 0.01204179 0.017478077463698855 0 +545 0 -3.437208 0.013325897 0.019354451635261342 0 +546 1 8.759242 0.99989384 0.00015315868008381413 1 +547 0 -4.5777693 0.0038230156 0.0055260154209878284 0 +548 0 -4.300773 0.005182322 0.0074959504499083774 0 +549 1 4.5371666 0.9889325 0.016056056373423 1 +550 0 -3.9912 0.007276599 0.010536293898534796 0 +551 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +552 0 -2.923206 0.02325764 0.033950027984242986 0 +553 0 -1.354846 0.11842905 0.18185141041152389 0 +554 0 -4.163023 0.006027712 0.0087224649837247988 0 +555 0 -1.7657175 0.07866246 0.11819830185137471 0 +556 0 -2.6747835 0.030367855 0.044490567396167725 0 +557 0 -3.069314 0.019864162 0.02894638789646586 0 +558 0 -4.3108807 0.005125152 0.0074130440548942798 0 +559 0 -3.437208 0.013325897 0.019354451635261342 0 +560 0 -3.2653847 0.01606247 0.023361373063565786 0 +561 0 -3.2653847 0.01606247 0.023361373063565786 0 +562 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +563 0 -3.9912 0.007276599 0.010536293898534796 0 +564 0 -3.3780956 0.014211106 0.0206493675057071 0 +565 1 8.5393095 0.9998647 0.00019521351546971681 1 +566 0 -3.4869633 0.0126231415 0.018327262393915984 0 +567 0 -3.09616 0.019295674 0.028109852883988089 0 +568 1 3.691854 0.9723495 0.040453091120168969 1 +569 1 7.642124 0.999636 0.00052524462619384972 1 +570 1 5.45413 0.9959468 0.0058593787035164354 1 +571 1 8.001436 0.9997551 0.00035338163228127936 1 +572 0 -3.9912 0.007276599 0.010536293898534796 0 +573 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +574 1 5.27071 0.9950423 0.007170202978073986 1 +575 0 -2.9279583 0.023138843 0.033774571618875787 0 +576 0 -3.437208 0.013325897 0.019354451635261342 0 +577 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +578 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +579 0 -4.5614243 0.0038923053 0.0056263664493250874 0 +580 0 -3.1047945 0.019116241 0.027845917617418135 0 +581 1 6.644459 0.9989066 0.0015782878712329485 1 +582 1 6.2932367 0.99838996 0.0023246695914181923 1 +583 0 -4.163023 0.006027712 0.0087224649837247988 0 +584 0 -2.3569767 0.042577 0.062771628508551322 0 +585 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +586 1 10.1500225 0.9999771 3.3021046862992363E-05 1 +587 0 -3.272923 0.015931563 0.023169444327594324 0 +588 1 4.3938894 0.98706186 0.018787595120546174 1 +589 0 -3.6090312 0.01105032 0.016030979963451743 0 +590 1 2.1874108 0.86993545 0.20101973436364209 1 +591 1 4.0009575 0.98018056 0.028880558672538456 1 +592 1 4.0920124 0.9820407 0.026145272763615762 1 +593 0 -3.2766178 0.015867786 0.023075947032752241 0 +594 1 3.4168415 0.9629125 0.054523388915238578 1 +595 0 -3.437208 0.013325897 0.019354451635261342 0 +596 0 -3.3817904 0.014154118 0.020565967564274219 0 +597 0 -2.6891708 0.029903961 0.043800514977400021 0 +598 0 -3.9912 0.007276599 0.010536293898534796 0 +599 0 -2.5371532 0.035171986 0.051656297504915218 0 +600 0 -3.9912 0.007276599 0.010536293898534796 0 +601 0 -4.704957 0.0033241964 0.0048037905231660839 0 +602 0 -3.7142038 0.009851713 0.014283491630889444 0 +603 1 3.307713 0.9583652 0.061352569779738138 1 +604 1 2.9647794 0.94036204 0.088711798400573361 1 +605 1 7.7185054 0.9996654 0.00048283608889884846 1 +606 0 -3.7483864 0.009490584 0.013757405165506545 0 +607 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +608 1 7.803549 0.99969536 0.00043956862262736329 1 +609 0 -3.6090312 0.01105032 0.016030979963451743 0 +610 1 5.3112736 0.9952583 0.0068571381335734557 1 +611 1 5.0439653 0.9936423 0.0092015462913481467 1 +612 1 11.885635 0.9999966 4.9015138759550746E-06 1 +613 0 -3.8232942 0.00874442 0.012671012142436589 0 +614 0 -4.405946 0.0046172338 0.0066766860483063314 0 +615 0 -3.0885918 0.019454302 0.028343226981323345 0 +616 0 -3.9912 0.007276599 0.010536293898534796 0 +617 0 ? ? ? 0 +618 0 -3.7142038 0.009851713 0.014283491630889444 0 +619 0 -3.437208 0.013325897 0.019354451635261342 0 +620 0 -3.9912 0.007276599 0.010536293898534796 0 +621 0 -0.27938604 0.30556083 0.52607976356436403 0 +622 0 -1.8795394 0.07003008 0.10474403942399887 0 +623 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +624 0 -3.1539075 0.018126339 0.026390691499228139 0 +625 0 -2.4943914 0.036808424 0.054105319982851556 0 +626 1 4.1199784 0.9825768 0.025357937820544574 1 +627 0 -3.2984931 0.015495301 0.022530003149323411 0 +628 0 -4.4408646 0.0044435263 0.0064249385978414108 0 +629 0 -3.8860273 0.008164489 0.011827214924006168 0 +630 0 -2.458294 0.03824661 0.056261084033362288 0 +631 0 -3.437208 0.013325897 0.019354451635261342 0 +632 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +633 1 3.809505 0.9756332 0.035589236295194712 1 +634 0 -4.163023 0.006027712 0.0087224649837247988 0 +635 0 -3.2250643 0.016780812 0.024415024437972225 0 +636 1 7.4516153 0.9995509 0.00064809018347703937 1 +637 0 -1.9161327 0.06744619 0.10074111664192874 0 +638 0 -3.8860273 0.008164489 0.011827214924006168 0 +639 0 -3.069314 0.019864162 0.02894638789646586 0 +640 0 -3.4302611 0.013427038 0.019502345114418035 0 +641 0 -3.9912 0.007276599 0.010536293898534796 0 +642 0 -3.9912 0.007276599 0.010536293898534796 0 +643 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +644 0 -4.4408646 0.0044435263 0.0064249385978414108 0 +645 0 -3.9912 0.007276599 0.010536293898534796 0 +646 0 -4.2502556 0.0054777074 0.007924384853928686 0 +647 0 -4.3785825 0.0047580656 0.0068808203063086652 0 +648 1 8.495777 0.999858 0.0002048458793528877 1 +649 0 -3.9912 0.007276599 0.010536293898534796 0 +650 0 -2.9605665 0.022339527 0.032594568703649685 0 +651 0 -3.9764798 0.0073948484 0.010708152663296801 0 +652 0 -3.272923 0.015931563 0.023169444327594324 0 +653 0 -3.7142038 0.009851713 0.014283491630889444 0 +654 0 -3.8193765 0.008781962 0.012725653416849284 0 +655 0 -3.9912 0.007276599 0.010536293898534796 0 +656 0 -3.437208 0.013325897 0.019354451635261342 0 +657 0 -0.4224856 0.27312568 0.46022215458764909 0 +658 1 6.8514137 0.9991296 0.0012562781410332333 1 +659 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +660 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +661 0 -3.5423806 0.011883428 0.017246843094017798 0 +662 0 -4.0451236 0.0068592355 0.0099298797488508467 0 +663 0 -4.0451236 0.0068592355 0.0099298797488508467 0 +664 0 -3.520666 0.012168032 0.017662437408366802 0 +665 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +666 0 -2.6617477 0.030794183 0.045125030965306769 0 +667 0 -3.8193765 0.008781962 0.012725653416849284 0 +668 1 2.0417032 0.85064197 0.23337606374990208 1 +669 1 6.798136 0.99907696 0.0013322766309305321 1 +670 1 4.4233613 0.98747057 0.018190347367794086 1 +671 0 -3.471818 0.012833092 0.018634062420200333 0 +672 0 -4.180897 0.005910713 0.0085526577863213431 0 +673 0 -2.9305813 0.023073524 0.033678106944583668 0 +674 0 -4.6772075 0.0034271795 0.0049528669508363014 0 +675 0 -3.0559897 0.020152379 0.029370685732283856 0 +676 0 -4.3682995 0.0048120874 0.0069591320626229613 0 +677 0 -3.6090312 0.01105032 0.016030979963451743 0 +678 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +679 0 -4.4408646 0.0044435263 0.0064249385978414108 0 +680 1 12.114309 0.9999974 3.7836232804574336E-06 1 +681 1 7.529873 0.999588 0.00059449451188820618 1 +682 0 -3.1010995 0.019192826 0.027958563399165529 0 +683 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +684 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +685 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +686 0 -4.7929907 0.0030174674 0.0043598665120151139 0 +687 0 -3.6626296 0.010422478 0.015115365276159273 0 +688 0 -3.8860273 0.008164489 0.011827214924006168 0 +689 0 -3.6151924 0.010976288 0.015922983970064867 0 +690 0 -4.3785825 0.0047580656 0.0068808203063086652 0 +691 1 3.916964 0.97829807 0.031654002126526269 1 +692 0 -4.163023 0.006027712 0.0087224649837247988 0 +693 0 -3.8830423 0.008191198 0.011866066162447262 0 +694 0 -3.716251 0.009829708 0.014251430548210188 0 +695 0 -4.4408646 0.0044435263 0.0064249385978414108 0 +696 1 5.3014183 0.99520665 0.0069319633534266831 1 +697 1 3.0674176 0.94640434 0.079471408542772501 1 +698 1 3.9357042 0.97873265 0.03101327353736098 1 diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris-out.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris-out.txt index 87cb55abe8..947924d424 100644 --- a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris-out.txt +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris-out.txt @@ -44,15 +44,15 @@ TRUTH ||======================== Precision ||1.0000 |0.9375 |0.9038 | Accuracy(micro-avg): 0.946667 Accuracy(macro-avg): 0.946667 -Log-loss: 0.433392 -Log-loss reduction: 0.605510 +Log-loss: 0.433400 +Log-loss reduction: 0.605503 OVERALL RESULTS --------------------------------------- Accuracy(micro-avg): 0.946667 (0.0000) Accuracy(macro-avg): 0.946667 (0.0000) -Log-loss: 0.433392 (0.0000) -Log-loss reduction: 0.605510 (0.0000) +Log-loss: 0.433400 (0.0000) +Log-loss reduction: 0.605503 (0.0000) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris-rp.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris-rp.txt index a3d2f79fe5..488aae593f 100644 --- a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris-rp.txt +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris-rp.txt @@ -1,4 +1,4 @@ WeightedEnsembleMulticlass Accuracy(micro-avg) Accuracy(macro-avg) Log-loss Log-loss reduction /oc /bp /nm Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings -0.946667 0.946667 0.433392 0.60551 MultiAverage mlr{t-} 5 WeightedEnsembleMulticlass %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsembleMulticlass{bp=mlr{t-} nm=5 oc=MultiAverage tp=-} dout=%Output% data=%Data% out=%Output% seed=1 xf=Term{col=Label} /oc:MultiAverage;/bp:mlr{t-};/nm:5 +0.946667 0.946667 0.4334 0.605503 MultiAverage mlr{t-} 5 WeightedEnsembleMulticlass %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsembleMulticlass{bp=mlr{t-} nm=5 oc=MultiAverage tp=-} dout=%Output% data=%Data% out=%Output% seed=1 xf=Term{col=Label} /oc:MultiAverage;/bp:mlr{t-};/nm:5 diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris.txt index d9f16cd7c2..843eed7f66 100644 --- a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris.txt +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Average-TrainTest-iris.txt @@ -1,151 +1,151 @@ Instance Label Assigned Log-loss #1 Score #2 Score #3 Score #1 Class #2 Class #3 Class -0 0 0 0.1995347399327862 0.81911176 0.16947956 0.011408707 0 1 2 -1 0 0 0.22944163401069054 0.79497737 0.19319294 0.011829739 0 1 2 -2 0 0 0.20840244858665047 0.81188023 0.17716658 0.0109532075 0 1 2 -3 0 0 0.23148277745006318 0.79335636 0.19420494 0.012438736 0 1 2 -4 0 0 0.19389359685209501 0.82374555 0.16494733 0.01130717 0 1 2 -5 0 0 0.25602002069266661 0.77412647 0.20466755 0.021206044 0 1 2 -6 0 0 0.2306506863851141 0.7940168 0.19150472 0.014478475 0 1 2 -7 0 0 0.21326688271208111 0.8079405 0.17985189 0.012207716 0 1 2 -8 0 0 0.23537507529361423 0.7902744 0.19786206 0.011863577 0 1 2 -9 0 0 0.2059855213161193 0.81384486 0.17629413 0.009860988 0 1 2 -10 0 0 0.19650580384489832 0.82159656 0.1664384 0.011965177 0 1 2 -11 0 0 0.22147755951658693 0.8013339 0.18572126 0.012944865 0 1 2 -12 0 0 0.20377273133806445 0.8156477 0.17500915 0.009343221 0 1 2 -13 0 0 0.18081681390014581 0.8345882 0.15767181 0.0077399695 0 1 2 -14 0 0 0.16075720855189568 0.8514988 0.13877651 0.009724712 0 1 2 -15 0 0 0.20729412091966262 0.81278056 0.16903628 0.018183162 0 1 2 -16 0 0 0.21900979288305267 0.80331385 0.17997597 0.016710162 0 1 2 -17 0 0 0.22495216027283546 0.7985544 0.18701091 0.014434678 0 1 2 -18 0 0 0.2335565532640983 0.7917128 0.19129768 0.016989535 0 1 2 -19 0 0 0.2153080656759096 0.806293 0.17871873 0.014988273 0 1 2 -20 0 0 0.23096331521848035 0.7937686 0.19241396 0.013817437 0 1 2 -21 0 0 0.24959942221453638 0.7791128 0.20183216 0.019055037 0 1 2 -22 0 0 0.16523656566316319 0.84769315 0.1435049 0.008802018 0 1 2 -23 0 0 0.33727896509923982 0.7137097 0.25882468 0.02746568 0 1 2 -24 0 0 0.24868295112435151 0.7798272 0.20468676 0.015486002 0 1 2 -25 0 0 0.24787629539957912 0.7804565 0.20620346 0.013340117 0 1 2 -26 0 0 0.28109516869097023 0.7549565 0.22445743 0.020586135 0 1 2 -27 0 0 0.207598284266892 0.8125334 0.17533147 0.012135103 0 1 2 -28 0 0 0.20534028005767088 0.81437016 0.17412049 0.011509328 0 1 2 -29 0 0 0.2340644842080086 0.7913108 0.19556604 0.013123191 0 1 2 -30 0 0 0.24081128540522556 0.78598994 0.20078325 0.013226831 0 1 2 -31 0 0 0.27111471805594867 0.762529 0.21798526 0.019485744 0 1 2 -32 0 0 0.15529321819210176 0.8561641 0.13475163 0.009084328 0 1 2 -33 0 0 0.16429901734220378 0.8484883 0.1407213 0.010790507 0 1 2 -34 0 0 0.2059855213161193 0.81384486 0.17629413 0.009860988 0 1 2 -35 0 0 0.20076723315534722 0.81810284 0.17156358 0.010333608 0 1 2 -36 0 0 0.1922879583805806 0.82506925 0.16416083 0.010770038 0 1 2 -37 0 0 0.2059855213161193 0.81384486 0.17629413 0.009860988 0 1 2 -38 0 0 0.22018274187250014 0.80237216 0.18653594 0.011091883 0 1 2 -39 0 0 0.21338145961265487 0.8078479 0.17993386 0.012218325 0 1 2 -40 0 0 0.21624608735187567 0.80553705 0.18088369 0.013579297 0 1 2 -41 0 0 0.30090695435667597 0.74014664 0.2452245 0.014628872 0 1 2 -42 0 0 0.2080697111998088 0.8121504 0.17692481 0.010924849 0 1 2 -43 0 0 0.346583722689219 0.7070996 0.2608438 0.032056604 0 1 2 -44 0 0 0.28356813015038723 0.7530918 0.22298017 0.023928061 0 1 2 -45 0 0 0.25792702699646886 0.7726516 0.21243982 0.014908629 0 1 2 -46 0 0 0.19841976452676924 0.82002556 0.16739146 0.012582937 0 1 2 -47 0 0 0.21653714682186442 0.8053026 0.18307057 0.011626892 0 1 2 -48 0 0 0.19639800430898482 0.82168514 0.16636033 0.011954611 0 1 2 -49 0 0 0.21097471142350205 0.80979455 0.1786267 0.01157874 0 1 2 -50 1 1 0.72280073006483325 0.4853909 0.39585024 0.1187588 1 2 0 -51 1 1 0.75095120185317177 0.47191745 0.4169356 0.111146964 1 2 0 -52 1 1 0.77019012421610067 0.46292505 0.44817072 0.08890428 1 2 0 -53 1 1 0.57142170157377936 0.564722 0.27788973 0.15738833 1 2 0 -54 1 1 0.71239956520915249 0.49046588 0.4126091 0.09692501 1 2 0 -55 1 1 0.63475024085845666 0.53006786 0.32890567 0.14102645 1 2 0 -56 1 2 0.82745693803480713 0.47644606 0.4371596 0.08639438 2 1 0 -57 1 1 0.60893856307030114 0.5439279 0.30701622 0.14905585 1 0 2 -58 1 1 0.64963193956024912 0.52223796 0.34073702 0.13702507 1 2 0 -59 1 1 0.63381208431825353 0.5305654 0.31399345 0.15544118 1 2 0 -60 1 1 0.54428602163438411 0.5802559 0.2640427 0.15570138 1 0 2 -61 1 1 0.70940332391985461 0.49193764 0.38475576 0.123306684 1 2 0 -62 1 1 0.53804079688154416 0.5838911 0.22874148 0.18736747 1 0 2 -63 1 1 0.6879192376376434 0.5026208 0.38552287 0.11185632 1 2 0 -64 1 1 0.6400093702291928 0.5272875 0.26100343 0.21170907 1 2 0 -65 1 1 0.69585737753817489 0.49864674 0.36727938 0.1340739 1 2 0 -66 1 1 0.72690633660133497 0.48340216 0.40959644 0.1070014 1 2 0 -67 1 1 0.59085655296881212 0.5538527 0.245508 0.20063941 1 0 2 -68 1 1 0.64115557553219493 0.52668345 0.38411126 0.08920532 1 2 0 -69 1 1 0.5749363545422802 0.5627407 0.22431958 0.21293981 1 0 2 -70 1 2 0.94975261920255172 0.55785084 0.3868367 0.055312466 2 1 0 -71 1 1 0.62600890570944601 0.5347217 0.29016036 0.175118 1 2 0 -72 1 1 0.70250976124137288 0.49534056 0.42644 0.078219526 1 2 0 -73 1 1 0.61466942923271062 0.54081964 0.30954513 0.14963524 1 2 0 -74 1 1 0.64122925176637446 0.52664465 0.31644267 0.15691268 1 2 0 -75 1 1 0.68437217417595786 0.5044068 0.36432958 0.13126366 1 2 0 -76 1 1 0.68371561310761098 0.5047381 0.39147708 0.1037849 1 2 0 -77 1 2 0.88400151338475097 0.5295436 0.41312647 0.05732989 2 1 0 -78 1 1 0.71607086970050315 0.48866853 0.4069286 0.104402855 1 2 0 -79 1 1 0.61714056045559618 0.53948486 0.29775333 0.16276182 1 0 2 -80 1 1 0.56736638067214262 0.5670168 0.22821276 0.20477043 1 0 2 -81 1 1 0.57782266402365767 0.5611188 0.26693997 0.17194131 1 0 2 -82 1 1 0.60233825031974264 0.5475299 0.24703716 0.20543304 1 2 0 -83 1 2 0.79663312659522156 0.48736793 0.45084435 0.061787795 2 1 0 -84 1 1 0.72666586419426138 0.48351842 0.40933466 0.10714693 1 2 0 -85 1 2 0.82134007875793547 0.46225557 0.43984184 0.097902596 2 1 0 -86 1 1 0.75394179064401501 0.47050825 0.4311025 0.09838925 1 2 0 -87 1 1 0.57947561020607474 0.56019205 0.30843455 0.13137347 1 2 0 -88 1 1 0.64942321187796115 0.522347 0.30190894 0.17574409 1 2 0 -89 1 1 0.59264806316544294 0.55286133 0.28258952 0.16454916 1 2 0 -90 1 1 0.58849175026178779 0.555164 0.2811874 0.16364865 1 2 0 -91 1 1 0.69355788128380058 0.4997947 0.3802913 0.11991403 1 2 0 -92 1 1 0.58944720965663988 0.5546338 0.25229675 0.1930695 1 2 0 -93 1 1 0.59601841622955909 0.55100113 0.3007882 0.14821072 1 0 2 -94 1 1 0.61682053844790008 0.53965753 0.3026878 0.15765478 1 2 0 -95 1 1 0.63319057823429603 0.53089523 0.27506185 0.19404295 1 2 0 -96 1 1 0.63915390942847394 0.52773875 0.30764136 0.16461992 1 2 0 -97 1 1 0.64112151201164758 0.5267014 0.3161953 0.15710333 1 2 0 -98 1 1 0.63046071020020678 0.5323465 0.31035924 0.15729432 1 0 2 -99 1 1 0.62664693219770329 0.5343806 0.29743573 0.16818376 1 2 0 -100 2 2 0.17846706690757469 0.8365516 0.1575896 0.005858842 2 1 0 -101 2 2 0.51421816746098636 0.5979679 0.36742076 0.03461136 2 1 0 -102 2 2 0.31769228863039173 0.7278267 0.25737938 0.014793965 2 1 0 -103 2 2 0.49804364886014763 0.6077184 0.3591608 0.033120845 2 1 0 -104 2 2 0.28854106605737767 0.74935603 0.23800834 0.012635595 2 1 0 -105 2 2 0.26617731127062982 0.76630324 0.22412863 0.009568109 2 1 0 -106 2 2 0.75896215077775475 0.46815205 0.4638621 0.067985885 2 1 0 -107 2 2 0.41923514123850858 0.65754956 0.32055864 0.021891778 2 1 0 -108 2 2 0.50003359845350104 0.6065103 0.36637592 0.02711381 2 1 0 -109 2 2 0.16468742397144318 0.8481588 0.14610842 0.0057328637 2 1 0 -110 2 2 0.42686799854970148 0.6525497 0.31649134 0.030959023 2 1 0 -111 2 2 0.48909386196347748 0.61318177 0.35608932 0.030728921 2 1 0 -112 2 2 0.35146201094322715 0.7036586 0.27744082 0.018900614 2 1 0 -113 2 2 0.48183973879425801 0.61764604 0.3536546 0.028699387 2 1 0 -114 2 2 0.27763969265791344 0.75756973 0.23042066 0.012009688 2 1 0 -115 2 2 0.28045847290187037 0.7554373 0.23031585 0.014246781 2 1 0 -116 2 2 0.50326655463589287 0.6045526 0.3597157 0.035731744 2 1 0 -117 2 2 0.20065380099943872 0.81819564 0.17359108 0.008213263 2 1 0 -118 2 2 0.20533801113624769 0.814372 0.18099868 0.004629333 2 1 0 -119 2 1 0.8600674942549501 0.5071842 0.42313352 0.0696823 1 2 0 -120 2 2 0.25264099171157628 0.7767467 0.21218887 0.011064478 2 1 0 -121 2 2 0.47450804449247858 0.6221911 0.34547982 0.032329112 2 1 0 -122 2 2 0.30311472728012528 0.73851436 0.25062612 0.010859603 2 1 0 -123 2 2 0.60313270674399444 0.54709506 0.4058976 0.04700739 2 1 0 -124 2 2 0.31989463066913293 0.72622555 0.25603 0.017744455 2 1 0 -125 2 2 0.43277646730676428 0.6487055 0.32355306 0.027741486 2 1 0 -126 2 2 0.61056267934985153 0.5430452 0.4062039 0.05075092 2 1 0 -127 2 2 0.58257378273067639 0.55845916 0.39139625 0.0501446 2 1 0 -128 2 2 0.35362581289669975 0.70213765 0.28067732 0.017185047 2 1 0 -129 2 2 0.58761507785171063 0.5556509 0.399575 0.044774093 2 1 0 -130 2 2 0.3963354248833964 0.672781 0.30772558 0.01949347 2 1 0 -131 2 2 0.27992082927769285 0.7558436 0.22843683 0.015719669 2 1 0 -132 2 2 0.31299549517404313 0.7312532 0.2549706 0.013776274 2 1 0 -133 2 1 0.79102632137371232 0.47108936 0.45337924 0.07553142 1 2 0 -134 2 1 0.80257605699684254 0.48491177 0.44817296 0.06691528 1 2 0 -135 2 2 0.23488223385069459 0.79066396 0.20108826 0.008247815 2 1 0 -136 2 2 0.22160230554755428 0.80123395 0.18902867 0.009737432 2 1 0 -137 2 2 0.4970530475312393 0.6083207 0.35517818 0.03650119 2 1 0 -138 2 2 0.59700697675868586 0.5504567 0.39653087 0.05301248 2 1 0 -139 2 2 0.35503113941059078 0.7011516 0.2784054 0.020443017 2 1 0 -140 2 2 0.23220578670909561 0.79278296 0.1980155 0.009201515 2 1 0 -141 2 2 0.29910573318000316 0.741481 0.24276999 0.01574904 2 1 0 -142 2 2 0.51421816746098636 0.5979679 0.36742076 0.03461136 2 1 0 -143 2 2 0.24011794940803582 0.7865351 0.20369373 0.009771238 2 1 0 -144 2 2 0.19264771595597938 0.8247725 0.16815351 0.0070741097 2 1 0 -145 2 2 0.29628142651594708 0.74357814 0.24185632 0.014565515 2 1 0 -146 2 2 0.5392498695980612 0.58318555 0.38181755 0.03499694 2 1 0 -147 2 2 0.42775366915807211 0.651972 0.3199589 0.028069088 2 1 0 -148 2 2 0.26528461622723248 0.7669876 0.21909939 0.013913065 2 1 0 -149 2 2 0.55556054938948174 0.57375056 0.38128686 0.044962626 2 1 0 +0 0 0 0.19970328346599298 0.8189737 0.16959019 0.011436053 0 1 2 +1 0 0 0.22961499477524977 0.79483956 0.193306 0.011854466 0 1 2 +2 0 0 0.20856507728348483 0.8117482 0.17727304 0.010978761 0 1 2 +3 0 0 0.23166851537753799 0.793209 0.19432405 0.012467007 0 1 2 +4 0 0 0.19406191590668695 0.8236069 0.16505754 0.011335572 0 1 2 +5 0 0 0.25627906880013401 0.77392596 0.2048155 0.021258527 0 1 2 +6 0 0 0.23084798254252556 0.79386014 0.1916256 0.014514317 0 1 2 +7 0 0 0.21344771805454324 0.8077944 0.1799692 0.012236389 0 1 2 +8 0 0 0.23555225896305373 0.7901344 0.19797596 0.011889741 0 1 2 +9 0 0 0.20614673189142804 0.81371367 0.17640372 0.009882608 0 1 2 +10 0 0 0.19668160150000139 0.82145214 0.16655372 0.011994141 0 1 2 +11 0 0 0.22167104525196937 0.8011789 0.18584502 0.012976128 0 1 2 +12 0 0 0.20392554588300318 0.8155231 0.17511351 0.009363453 0 1 2 +13 0 0 0.18094501747645839 0.83448124 0.1577603 0.007758487 0 1 2 +14 0 0 0.16089854789960872 0.85137844 0.13887243 0.009749197 0 1 2 +15 0 0 0.20751796207141654 0.81259865 0.169168 0.018233404 0 1 2 +16 0 0 0.21921363680728581 0.8031501 0.18009754 0.016752362 0 1 2 +17 0 0 0.22514527435671869 0.7984002 0.18713103 0.014468761 0 1 2 +18 0 0 0.23378055327346786 0.7915355 0.19143555 0.017029036 0 1 2 +19 0 0 0.21551100846533355 0.8061294 0.17884374 0.015026912 0 1 2 +20 0 0 0.23116457856721909 0.79360884 0.19254373 0.013847442 0 1 2 +21 0 0 0.24983347297432063 0.7789305 0.20196758 0.019101946 0 1 2 +22 0 0 0.16536897557758942 0.8475809 0.14359327 0.008825864 0 1 2 +23 0 0 0.3375820827910338 0.7134934 0.25898376 0.027522791 0 1 2 +24 0 0 0.24891426457605365 0.7796468 0.20483045 0.015522817 0 1 2 +25 0 0 0.24807197836813213 0.7803038 0.20632899 0.0133671975 0 1 2 +26 0 0 0.2813463436815471 0.7547669 0.22460055 0.020632578 0 1 2 +27 0 0 0.20777677667995276 0.81238836 0.1754481 0.012163586 0 1 2 +28 0 0 0.2055089265644528 0.8142328 0.17423156 0.011535589 0 1 2 +29 0 0 0.23426042096543317 0.79115576 0.19569093 0.013153376 0 1 2 +30 0 0 0.24100748682652162 0.78583574 0.20090853 0.01325572 0 1 2 +31 0 0 0.27134768300343914 0.7623514 0.2181216 0.019527063 0 1 2 +32 0 0 0.15544353528730609 0.8560354 0.13485458 0.009110014 0 1 2 +33 0 0 0.16445947675667871 0.84835213 0.14082752 0.01082035 0 1 2 +34 0 0 0.20614673189142804 0.81371367 0.17640372 0.009882608 0 1 2 +35 0 0 0.20091805883881336 0.81797945 0.17166401 0.010356615 0 1 2 +36 0 0 0.19244401325586014 0.8249405 0.16426522 0.010794281 0 1 2 +37 0 0 0.20614673189142804 0.81371367 0.17640372 0.009882608 0 1 2 +38 0 0 0.22034841233753494 0.80223924 0.18664348 0.0111173345 0 1 2 +39 0 0 0.21356150392913109 0.8077025 0.18005112 0.012246549 0 1 2 +40 0 0 0.21642820236329943 0.80539036 0.18099771 0.013612032 0 1 2 +41 0 0 0.30109718633959498 0.74000585 0.24534102 0.014653136 0 1 2 +42 0 0 0.2082341208670859 0.8120169 0.17703156 0.010951618 0 1 2 +43 0 0 0.34691109201072423 0.7068682 0.2610038 0.032128103 0 1 2 +44 0 0 0.28386275713631315 0.75286996 0.22314262 0.023987444 0 1 2 +45 0 0 0.25812530410024825 0.7724984 0.21256189 0.014939775 0 1 2 +46 0 0 0.19860738553187271 0.8198717 0.16751266 0.012615676 0 1 2 +47 0 0 0.21671102367241341 0.8051626 0.18318307 0.011654347 0 1 2 +48 0 0 0.19657429087856665 0.8215403 0.16647571 0.011984015 0 1 2 +49 0 0 0.21114526794350838 0.80965644 0.17873843 0.011605174 0 1 2 +50 1 1 0.72247991244247844 0.48554665 0.39589342 0.11855998 1 2 0 +51 1 1 0.75072457659184566 0.4720244 0.4170204 0.1109552 1 2 0 +52 1 1 0.7698654528384179 0.46307537 0.4481743 0.088750415 1 2 0 +53 1 1 0.57112916841714423 0.5648872 0.27793488 0.15717794 1 2 0 +54 1 1 0.71208218813117408 0.49062157 0.41260916 0.096769296 1 2 0 +55 1 1 0.6344920954183777 0.5302047 0.32899427 0.14080106 1 2 0 +56 1 2 0.82727581974235642 0.4765273 0.43723878 0.086233966 2 1 0 +57 1 1 0.60857810364607623 0.544124 0.3067079 0.1491681 1 0 2 +58 1 1 0.6492959881075514 0.52241343 0.34077618 0.13681035 1 2 0 +59 1 1 0.633612360628237 0.53067136 0.31411797 0.15521072 1 2 0 +60 1 1 0.54393570063930774 0.58045924 0.2637764 0.15576443 1 0 2 +61 1 1 0.70920372778183449 0.49203584 0.38485917 0.123105004 1 2 0 +62 1 1 0.5376425530681963 0.5841237 0.22847982 0.18739656 1 0 2 +63 1 1 0.68765256946748587 0.50275487 0.38557705 0.11166804 1 2 0 +64 1 1 0.6397495247369005 0.5274245 0.26115075 0.21142475 1 2 0 +65 1 1 0.69556032286798197 0.49879488 0.36734512 0.13386004 1 2 0 +66 1 1 0.72673865940174309 0.48348323 0.40970302 0.106813766 1 2 0 +67 1 1 0.59047404332756848 0.5540646 0.24519639 0.20073903 1 0 2 +68 1 1 0.64077358735605827 0.5268847 0.3840363 0.08907913 1 2 0 +69 1 1 0.57459366102722864 0.56293356 0.22404075 0.21302576 1 0 2 +70 1 2 0.94961387782084783 0.5579045 0.38689038 0.055205192 2 1 0 +71 1 1 0.6257099917612241 0.53488153 0.2902473 0.17487113 1 2 0 +72 1 1 0.7021410759761163 0.4955232 0.42638198 0.07809481 1 2 0 +73 1 1 0.61434843411645212 0.5409933 0.30960724 0.14939952 1 2 0 +74 1 1 0.64091477925576956 0.5268103 0.3165122 0.15667751 1 2 0 +75 1 1 0.68407148232842552 0.5045585 0.3643849 0.13105662 1 2 0 +76 1 1 0.68334853791760064 0.5049234 0.39145765 0.10361897 1 2 0 +77 1 2 0.88366807271212022 0.529507 0.41326424 0.05722883 2 1 0 +78 1 1 0.7158362810587191 0.48878318 0.4069887 0.104228094 1 2 0 +79 1 1 0.61674355844515727 0.5396991 0.2974386 0.16286233 1 0 2 +80 1 1 0.56702543043296505 0.56721014 0.22793916 0.2048507 1 0 2 +81 1 1 0.57744553142383837 0.56133044 0.26664442 0.17202519 1 0 2 +82 1 1 0.6020230385750599 0.5477025 0.24713679 0.20516074 1 2 0 +83 1 2 0.79632572715929228 0.48733798 0.45098296 0.061679114 2 1 0 +84 1 1 0.72652207687105719 0.48358795 0.40945554 0.10695656 1 2 0 +85 1 2 0.82122862485727066 0.46239072 0.43989086 0.09771848 2 1 0 +86 1 1 0.75365249217115637 0.47064438 0.43113437 0.09822126 1 2 0 +87 1 1 0.57910423546049083 0.5604001 0.30840737 0.13119255 1 2 0 +88 1 1 0.64918954334483836 0.52246904 0.30205613 0.17547482 1 2 0 +89 1 1 0.59237350572858261 0.55301315 0.28266603 0.1643209 1 2 0 +90 1 1 0.58820287537271632 0.5553244 0.2812703 0.16340537 1 2 0 +91 1 1 0.69330931833735343 0.49991894 0.3803693 0.11971181 1 2 0 +92 1 1 0.58912829945004608 0.5548107 0.25237796 0.1928114 1 2 0 +93 1 1 0.5956565275778597 0.55120057 0.30049133 0.14830817 1 0 2 +94 1 1 0.61655836684835097 0.53979903 0.30278113 0.15741988 1 2 0 +95 1 1 0.63291139686141251 0.53104347 0.2752038 0.19375281 1 2 0 +96 1 1 0.63890354510303216 0.5278709 0.30776128 0.16436782 1 2 0 +97 1 1 0.64082585318732821 0.52685714 0.31627795 0.15686497 1 2 0 +98 1 1 0.63011893676892994 0.53252846 0.31005707 0.15741442 1 0 2 +99 1 1 0.62638462526698035 0.5345208 0.29754353 0.16793567 1 2 0 +100 2 2 0.17854573045386624 0.8364858 0.15766709 0.0058471747 2 1 0 +101 2 2 0.51430867978572659 0.5979138 0.3675367 0.034549545 2 1 0 +102 2 2 0.31788033494603257 0.72768986 0.25754195 0.0147682 2 1 0 +103 2 2 0.49816605941844655 0.607644 0.35929748 0.033058584 2 1 0 +104 2 2 0.28867732934052415 0.7492539 0.23813371 0.012612431 2 1 0 +105 2 2 0.26640780607662956 0.76612663 0.22432148 0.009551965 2 1 0 +106 2 2 0.75887424088069566 0.4681932 0.46393743 0.06786944 2 1 0 +107 2 2 0.41949732562943226 0.6573772 0.3207701 0.021852715 2 1 0 +108 2 2 0.50032345268755207 0.6063345 0.3665964 0.027069187 2 1 0 +109 2 2 0.16477470975062117 0.84808475 0.14619362 0.005721667 2 1 0 +110 2 2 0.42691933359896422 0.6525162 0.3165834 0.030900424 2 1 0 +111 2 2 0.4892665138472555 0.6130759 0.35624778 0.030676395 2 1 0 +112 2 2 0.35161356283616868 0.70355195 0.2775808 0.018867258 2 1 0 +113 2 2 0.48197340426799445 0.6175635 0.35378543 0.028651152 2 1 0 +114 2 2 0.27773285264554587 0.75749916 0.23051243 0.011988522 2 1 0 +115 2 2 0.28053556200171892 0.7553791 0.23040111 0.014219812 2 1 0 +116 2 2 0.50337609743580891 0.6044864 0.35984895 0.03566469 2 1 0 +117 2 2 0.20076104031638917 0.8181079 0.17369609 0.008196038 2 1 0 +118 2 2 0.20559582282446537 0.8141621 0.18121485 0.00462307 2 1 0 +119 2 1 0.86029078998818231 0.50738615 0.42303905 0.06957479 1 2 0 +120 2 2 0.25276446799502034 0.7766508 0.2123049 0.01104436 2 1 0 +121 2 2 0.47454626861507559 0.6221673 0.34556285 0.032269914 2 1 0 +122 2 2 0.30340661307510058 0.73829883 0.250859 0.010842184 2 1 0 +123 2 2 0.60325288306884739 0.5470293 0.40604225 0.04692855 2 1 0 +124 2 2 0.31998820004798173 0.7261576 0.2561329 0.017709536 2 1 0 +125 2 2 0.43294150179254648 0.64859843 0.32371324 0.02768834 2 1 0 +126 2 2 0.61062798868369261 0.54300976 0.4063266 0.050663676 2 1 0 +127 2 2 0.58257581061288266 0.55845803 0.3914896 0.050052296 2 1 0 +128 2 2 0.3537892399929789 0.7020229 0.28082222 0.017154936 2 1 0 +129 2 2 0.58781333237576538 0.55554074 0.39976707 0.044692274 2 1 0 +130 2 2 0.39661586555721279 0.67259234 0.3079466 0.0194611 2 1 0 +131 2 2 0.28003683676996688 0.7557559 0.22855721 0.01568691 2 1 0 +132 2 2 0.31315469739923252 0.7311368 0.2551107 0.0137525005 2 1 0 +133 2 1 0.79106931217684229 0.47124082 0.45335975 0.075399496 1 2 0 +134 2 1 0.80270726491286482 0.48509088 0.44811416 0.06679502 1 2 0 +135 2 2 0.23508888702042546 0.7905006 0.20126441 0.008235063 2 1 0 +136 2 2 0.22165847237718772 0.80118895 0.18909346 0.009717573 2 1 0 +137 2 2 0.49712418518998785 0.60827744 0.3552923 0.036430355 2 1 0 +138 2 2 0.59698402120269867 0.55046934 0.39661556 0.052915093 2 1 0 +139 2 2 0.35516605880793273 0.701057 0.27853647 0.02040651 2 1 0 +140 2 2 0.23232518613015027 0.7926883 0.19812645 0.00918522 2 1 0 +141 2 2 0.29922712331098988 0.741391 0.24288689 0.015722154 2 1 0 +142 2 2 0.51430867978572659 0.5979138 0.3675367 0.034549545 2 1 0 +143 2 2 0.24024026789369451 0.7864389 0.20380802 0.009753074 2 1 0 +144 2 2 0.19273906686103115 0.82469714 0.16824202 0.0070608924 2 1 0 +145 2 2 0.29640936881831181 0.743483 0.24197637 0.014540671 2 1 0 +146 2 2 0.53944786083097751 0.5830701 0.38198867 0.034941282 2 1 0 +147 2 2 0.42785935868312264 0.6519031 0.3200786 0.028018346 2 1 0 +148 2 2 0.26532168585094545 0.7669592 0.21915643 0.013884395 2 1 0 +149 2 2 0.55556054938948174 0.57375056 0.38137275 0.044876736 2 1 0 diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris-out.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris-out.txt index 690fc87b19..dad57accee 100644 --- a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris-out.txt +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris-out.txt @@ -41,25 +41,25 @@ Beginning training model 7 of 20 Beginning optimization num vars: 15 improvement criterion: Mean Improvement -L1 regularization selected 12 of 15 weights. +L1 regularization selected 11 of 15 weights. Trainer 7 of 20 finished in %Time% Beginning training model 8 of 20 Beginning optimization num vars: 15 improvement criterion: Mean Improvement -L1 regularization selected 13 of 15 weights. +L1 regularization selected 11 of 15 weights. Trainer 8 of 20 finished in %Time% Beginning training model 9 of 20 Beginning optimization num vars: 15 improvement criterion: Mean Improvement -L1 regularization selected 12 of 15 weights. +L1 regularization selected 11 of 15 weights. Trainer 9 of 20 finished in %Time% Beginning training model 10 of 20 Beginning optimization num vars: 15 improvement criterion: Mean Improvement -L1 regularization selected 11 of 15 weights. +L1 regularization selected 12 of 15 weights. Trainer 10 of 20 finished in %Time% Beginning training model 11 of 20 Beginning optimization @@ -77,7 +77,7 @@ Beginning training model 13 of 20 Beginning optimization num vars: 15 improvement criterion: Mean Improvement -L1 regularization selected 10 of 15 weights. +L1 regularization selected 12 of 15 weights. Trainer 13 of 20 finished in %Time% Beginning training model 14 of 20 Beginning optimization @@ -101,7 +101,7 @@ Beginning training model 17 of 20 Beginning optimization num vars: 15 improvement criterion: Mean Improvement -L1 regularization selected 11 of 15 weights. +L1 regularization selected 10 of 15 weights. Trainer 17 of 20 finished in %Time% Beginning training model 18 of 20 Beginning optimization @@ -134,15 +134,15 @@ TRUTH ||======================== Precision ||1.0000 |0.9184 |0.9020 | Accuracy(micro-avg): 0.940000 Accuracy(macro-avg): 0.940000 -Log-loss: 0.435613 -Log-loss reduction: 0.603488 +Log-loss: 0.435078 +Log-loss reduction: 0.603975 OVERALL RESULTS --------------------------------------- Accuracy(micro-avg): 0.940000 (0.0000) Accuracy(macro-avg): 0.940000 (0.0000) -Log-loss: 0.435613 (0.0000) -Log-loss reduction: 0.603488 (0.0000) +Log-loss: 0.435078 (0.0000) +Log-loss reduction: 0.603975 (0.0000) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris-rp.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris-rp.txt index 7948bf23ec..f02650755c 100644 --- a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris-rp.txt +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris-rp.txt @@ -1,4 +1,4 @@ WeightedEnsembleMulticlass Accuracy(micro-avg) Accuracy(macro-avg) Log-loss Log-loss reduction /bp /nm Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings -0.94 0.94 0.435613 0.603488 mlr{t-} 20 WeightedEnsembleMulticlass %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsembleMulticlass{bp=mlr{t-} nm=20 st=BootstrapSelector{} tp=-} dout=%Output% data=%Data% out=%Output% seed=1 xf=Term{col=Label} /bp:mlr{t-};/nm:20 +0.94 0.94 0.435078 0.603975 mlr{t-} 20 WeightedEnsembleMulticlass %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsembleMulticlass{bp=mlr{t-} nm=20 st=BootstrapSelector{} tp=-} dout=%Output% data=%Data% out=%Output% seed=1 xf=Term{col=Label} /bp:mlr{t-};/nm:20 diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris.txt index d97c7e23c3..319212e3c8 100644 --- a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris.txt +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Bootstrap-TrainTest-iris.txt @@ -1,151 +1,151 @@ Instance Label Assigned Log-loss #1 Score #2 Score #3 Score #1 Class #2 Class #3 Class -0 0 0 0.19248787194951467 0.8249043 0.16391772 0.0112357875 0 1 2 -1 0 0 0.21089485356602972 0.8098592 0.17780837 0.011608226 0 1 2 -2 0 0 0.19409556864309002 0.8235792 0.16561954 0.010706931 0 1 2 -3 0 0 0.21419006839095034 0.80719495 0.18040752 0.012171622 0 1 2 -4 0 0 0.18833345903842305 0.82833844 0.16069299 0.011115085 0 1 2 -5 0 0 0.25671045116465646 0.7735922 0.20349255 0.020969568 0 1 2 -6 0 0 0.22057586389907957 0.8020568 0.18351358 0.014028646 0 1 2 -7 0 0 0.2040367914342095 0.81543237 0.17245781 0.011987341 0 1 2 -8 0 0 0.21469098821931784 0.8067907 0.1807341 0.011613437 0 1 2 -9 0 0 0.18990947648168735 0.827034 0.16333845 0.009730559 0 1 2 -10 0 0 0.19308170709676575 0.8244146 0.16429962 0.011793552 0 1 2 -11 0 0 0.21188952031136538 0.8090541 0.17814124 0.012651183 0 1 2 -12 0 0 0.18612689123641185 0.83016825 0.16011591 0.009217645 0 1 2 -13 0 0 0.16365933251526441 0.8490312 0.1433618 0.0075960676 0 1 2 -14 0 0 0.16113954990336041 0.8511733 0.13809414 0.00960567 0 1 2 -15 0 0 0.21012891014784502 0.81047976 0.16992986 0.017931463 0 1 2 -16 0 0 0.21873203245527259 0.803537 0.17837805 0.01636998 0 1 2 -17 0 0 0.21780768458867614 0.8042801 0.18159494 0.014109329 0 1 2 -18 0 0 0.23289034631477437 0.79224044 0.19024733 0.016867355 0 1 2 -19 0 0 0.21337843455473515 0.80785036 0.17651081 0.014678294 0 1 2 -20 0 0 0.22228998975816838 0.80068314 0.18572956 0.0136960875 0 1 2 -21 0 0 0.24632448530288342 0.78166854 0.19899875 0.018468987 0 1 2 -22 0 0 0.15908638273424053 0.8529227 0.13780464 0.0085659865 0 1 2 -23 0 0 0.32499574971623529 0.7225304 0.2499164 0.02643568 0 1 2 -24 0 0 0.23875897794580808 0.7876047 0.19739422 0.015246329 0 1 2 -25 0 0 0.22896249883151112 0.79535836 0.19072814 0.013115302 0 1 2 -26 0 0 0.27143494156468528 0.7622849 0.21734704 0.019897442 0 1 2 -27 0 0 0.20061861560427344 0.81822443 0.17000848 0.011955516 0 1 2 -28 0 0 0.19645364366816578 0.8216394 0.1669108 0.011345269 0 1 2 -29 0 0 0.21918239337856113 0.8031752 0.18391031 0.012841332 0 1 2 -30 0 0 0.22350078104580848 0.79971427 0.18718359 0.012965914 0 1 2 -31 0 0 0.26231798849203691 0.76926637 0.21132784 0.019094171 0 1 2 -32 0 0 0.15546038557804362 0.856021 0.13391201 0.009047328 0 1 2 -33 0 0 0.16523494844385578 0.8476945 0.14044563 0.010695541 0 1 2 -34 0 0 0.18990947648168735 0.827034 0.16333845 0.009730559 0 1 2 -35 0 0 0.18712876342819015 0.82933694 0.16043445 0.010163587 0 1 2 -36 0 0 0.18586687170692034 0.83038414 0.15897799 0.0106012 0 1 2 -37 0 0 0.18990947648168735 0.827034 0.16333845 0.009730559 0 1 2 -38 0 0 0.20150773594974408 0.81749725 0.17104076 0.010840238 0 1 2 -39 0 0 0.20432212486997534 0.81519973 0.17266783 0.012038765 0 1 2 -40 0 0 0.20901646458364931 0.8113819 0.17520803 0.013237229 0 1 2 -41 0 0 0.27330587248955635 0.76086 0.22409466 0.014374888 0 1 2 -42 0 0 0.19328185173953721 0.8242496 0.16500969 0.01065601 0 1 2 -43 0 0 0.33922807226712121 0.71231997 0.25515014 0.030492488 0 1 2 -44 0 0 0.28280339175118763 0.75366795 0.22098917 0.023443494 0 1 2 -45 0 0 0.23839851113284732 0.78788865 0.19637284 0.014528947 0 1 2 -46 0 0 0.19617540530110095 0.82186806 0.16539367 0.012385834 0 1 2 -47 0 0 0.20186513723566452 0.81720513 0.17134875 0.0113613885 0 1 2 -48 0 0 0.19283201629927799 0.8246205 0.16403821 0.011762289 0 1 2 -49 0 0 0.19942057437754529 0.8192053 0.16932516 0.011371488 0 1 2 -50 1 1 0.69241014278315771 0.50036865 0.38920426 0.11844209 1 2 0 -51 1 1 0.72255516581343215 0.4855101 0.41205072 0.11150342 1 2 0 -52 1 1 0.74117353476279446 0.47655433 0.44188732 0.089149326 1 2 0 -53 1 1 0.58850613714593147 0.555156 0.2788548 0.1635042 1 2 0 -54 1 1 0.69292428403176232 0.50011146 0.4086107 0.098359935 1 2 0 -55 1 1 0.63232735505457305 0.5313537 0.3208812 0.14350468 1 2 0 -56 1 2 0.79957813256768062 0.47135997 0.44951856 0.08678871 2 1 0 -57 1 1 0.64545404818887431 0.5244244 0.31833512 0.14914165 1 0 2 -58 1 1 0.63723447138971445 0.5287527 0.3334638 0.13977997 1 2 0 -59 1 1 0.64251816210224766 0.5259663 0.3096367 0.15934971 1 2 0 -60 1 1 0.58106958714568646 0.5592998 0.27867922 0.15798523 1 0 2 -61 1 1 0.69328499599477678 0.4999311 0.3815279 0.12497688 1 2 0 -62 1 1 0.5580943155087128 0.57229865 0.23871075 0.18836403 1 0 2 -63 1 1 0.66931338309259314 0.51206005 0.37796557 0.11335312 1 2 0 -64 1 1 0.64360156033662053 0.52539676 0.25747693 0.2162116 1 2 0 -65 1 1 0.67333124866058358 0.5100068 0.36244223 0.13508126 1 2 0 -66 1 1 0.70450122949046701 0.49435508 0.4036824 0.10834631 1 2 0 -67 1 1 0.59234321955604496 0.5530299 0.25144726 0.19728798 1 0 2 -68 1 1 0.6489437254839121 0.5225975 0.38229865 0.091786325 1 2 0 -69 1 1 0.58436589905432745 0.55745924 0.23120558 0.21169138 1 0 2 -70 1 2 0.93065522089585184 0.5504366 0.39429528 0.056838527 2 1 0 -71 1 1 0.6311201827267473 0.53199553 0.28625715 0.17891102 1 2 0 -72 1 1 0.68881426363582232 0.50217116 0.42115444 0.07984154 1 2 0 -73 1 1 0.61156715840902287 0.5425 0.2997402 0.15276724 1 2 0 -74 1 1 0.63636572455132645 0.52921224 0.3108666 0.16020674 1 2 0 -75 1 1 0.66665434964644166 0.51342344 0.3595778 0.13271306 1 2 0 -76 1 1 0.66442067116396375 0.51457155 0.3850018 0.105545744 1 2 0 -77 1 2 0.86434235863295594 0.5227423 0.42132854 0.05849034 2 1 0 -78 1 1 0.69656095603616708 0.49829602 0.40220338 0.10570199 1 2 0 -79 1 1 0.63943393549001015 0.527591 0.30639994 0.16354321 1 0 2 -80 1 1 0.58296598761567642 0.5582402 0.23609358 0.20516095 1 0 2 -81 1 1 0.60123663978445951 0.5481334 0.2763446 0.17248619 1 0 2 -82 1 1 0.60769908378196724 0.5446025 0.24416274 0.21020496 1 2 0 -83 1 2 0.77339944038260022 0.4813576 0.46144176 0.062465645 2 1 0 -84 1 1 0.70499864625313913 0.49410924 0.4029985 0.10868049 1 2 0 -85 1 2 0.79326093755798632 0.45757008 0.45236725 0.09825148 2 1 0 -86 1 1 0.72625230801289853 0.48371843 0.42579013 0.098693565 1 2 0 -87 1 1 0.59423544078506696 0.5519844 0.30595416 0.13506347 1 2 0 -88 1 1 0.648188399129075 0.5229924 0.29593158 0.17919323 1 2 0 -89 1 1 0.60550196008274937 0.5458004 0.28122658 0.16939 1 2 0 -90 1 1 0.59587769022701786 0.5510787 0.27543032 0.16759156 1 2 0 -91 1 1 0.6730458921119522 0.51015234 0.37311378 0.12094216 1 2 0 -92 1 1 0.5971153730828247 0.55039704 0.24922246 0.19817561 1 2 0 -93 1 1 0.63363370152628307 0.53066003 0.31326824 0.1488813 1 0 2 -94 1 1 0.62454850856523225 0.53550315 0.2968166 0.16116163 1 2 0 -95 1 1 0.63105811463354766 0.53202856 0.26759404 0.19787921 1 2 0 -96 1 1 0.63949686460061372 0.5275578 0.30142212 0.16823013 1 2 0 -97 1 1 0.63708748633030077 0.5288304 0.31023777 0.16045232 1 2 0 -98 1 1 0.66619101547997928 0.5136614 0.32096612 0.15678167 1 0 2 -99 1 1 0.63213196515632697 0.53145754 0.29214776 0.17133752 1 2 0 -100 2 2 0.19863348515953283 0.8198503 0.17448631 0.0062297145 2 1 0 -101 2 2 0.52501069302045333 0.59154904 0.37328643 0.03593976 2 1 0 -102 2 2 0.34381233702191072 0.709062 0.27672002 0.015025256 2 1 0 -103 2 2 0.52145273603867826 0.5936575 0.37406993 0.033773787 2 1 0 -104 2 2 0.31030398353558158 0.73322403 0.2546577 0.013094165 2 1 0 -105 2 2 0.29537627755747542 0.7442515 0.24621645 0.009779885 2 1 0 -106 2 1 0.76594573522387388 0.46913236 0.46489406 0.07052642 1 2 0 -107 2 2 0.44838283951888097 0.63866013 0.3395981 0.022044823 2 1 0 -108 2 2 0.51684805575507486 0.5963974 0.3758612 0.027623251 2 1 0 -109 2 2 0.18507445180289583 0.8310424 0.1626103 0.0059933364 2 1 0 -110 2 2 0.44600888450111886 0.6401781 0.32872945 0.03205818 2 1 0 -111 2 2 0.50306189655416156 0.60467637 0.3641055 0.03156344 2 1 0 -112 2 2 0.37179052091148973 0.68949866 0.29132533 0.01932842 2 1 0 -113 2 2 0.48954694225112294 0.612904 0.3577423 0.030060705 2 1 0 -114 2 2 0.2931546594710604 0.7459068 0.24135244 0.013048984 2 1 0 -115 2 2 0.30011066022158511 0.74073625 0.2447896 0.014945621 2 1 0 -116 2 2 0.52625969100020198 0.59081066 0.37459797 0.036467478 2 1 0 -117 2 2 0.22349593644982671 0.79971814 0.19123581 0.008397197 2 1 0 -118 2 2 0.22707749669640795 0.796859 0.19839063 0.0047730953 2 1 0 -119 2 1 0.87281946162373114 0.5088998 0.417772 0.071826816 1 2 0 -120 2 2 0.27443042724324229 0.7600049 0.22898 0.011382496 2 1 0 -121 2 2 0.48525210729891233 0.615542 0.35165885 0.03391049 2 1 0 -122 2 2 0.32944481565648143 0.719323 0.2696116 0.010995779 2 1 0 -123 2 2 0.61067442127485749 0.54298455 0.40980455 0.047920696 2 1 0 -124 2 2 0.3461925039189433 0.7073763 0.27530143 0.018277043 2 1 0 -125 2 2 0.46653750574697783 0.6271701 0.34503144 0.027875492 2 1 0 -126 2 2 0.61815355212612122 0.53893864 0.41050735 0.051708613 2 1 0 -127 2 2 0.59457931610341364 0.55179465 0.39782476 0.051486164 2 1 0 -128 2 2 0.37207202985661819 0.6893046 0.29315808 0.017891679 2 1 0 -129 2 2 0.61538992209199816 0.5404301 0.41530675 0.045012504 2 1 0 -130 2 2 0.42238653648190261 0.6554806 0.3249418 0.019594416 2 1 0 -131 2 2 0.30891827916972553 0.7342408 0.24857488 0.015740428 2 1 0 -132 2 2 0.33116178378377054 0.718089 0.26780963 0.014458399 2 1 0 -133 2 1 0.80836695796973468 0.48371834 0.44558513 0.07637462 1 2 0 -134 2 1 0.81852819427543022 0.4924503 0.44108036 0.06790538 1 2 0 -135 2 2 0.25950329679705686 0.77143466 0.22066763 0.008442265 2 1 0 -136 2 2 0.2420007640180428 0.7850556 0.20459294 0.010242797 2 1 0 -137 2 2 0.52173480619447854 0.59349006 0.3713804 0.0371928 2 1 0 -138 2 2 0.60748174736041938 0.5447209 0.4024183 0.05444832 2 1 0 -139 2 2 0.37582931351801802 0.68671954 0.29260468 0.0209805 2 1 0 -140 2 2 0.25180453260226143 0.7773967 0.21317714 0.009634437 2 1 0 -141 2 2 0.31734004266572607 0.72808313 0.25591937 0.016350273 2 1 0 -142 2 2 0.52501069302045333 0.59154904 0.37328643 0.03593976 2 1 0 -143 2 2 0.26364723070966672 0.7682445 0.22250286 0.010081337 2 1 0 -144 2 2 0.2126170750558016 0.80846566 0.1843186 0.007439901 2 1 0 -145 2 2 0.31398201133178771 0.73053217 0.25462818 0.015249444 2 1 0 -146 2 2 0.54649295701011613 0.57897675 0.38531333 0.036110558 2 1 0 -147 2 2 0.44538266109724217 0.6405791 0.33107612 0.02891146 2 1 0 -148 2 2 0.28676497894537284 0.75068814 0.23464048 0.014580958 2 1 0 -149 2 2 0.5703671057060149 0.56531787 0.38917458 0.046133474 2 1 0 +0 0 0 0.19152855003263652 0.82569605 0.16331035 0.011343198 0 1 2 +1 0 0 0.21086180826286643 0.809886 0.17782414 0.011701666 0 1 2 +2 0 0 0.19403955374687626 0.8236253 0.16563997 0.0107215755 0 1 2 +3 0 0 0.2141724203810971 0.8072092 0.18042812 0.0122675905 0 1 2 +4 0 0 0.18761810385746255 0.8289312 0.15999807 0.011208784 0 1 2 +5 0 0 0.25504826445842216 0.7748791 0.2022925 0.021144144 0 1 2 +6 0 0 0.22018623329872733 0.80236936 0.18347096 0.01407827 0 1 2 +7 0 0 0.20290130901732839 0.8163588 0.17211688 0.012109447 0 1 2 +8 0 0 0.21478984279921357 0.80671096 0.18079332 0.011648225 0 1 2 +9 0 0 0.18927891546323997 0.82755566 0.16317075 0.009776272 0 1 2 +10 0 0 0.19192471938992742 0.825369 0.16281055 0.011974884 0 1 2 +11 0 0 0.21076841886688111 0.8099616 0.17780253 0.012802368 0 1 2 +12 0 0 0.18600699531167428 0.8302678 0.1601362 0.009262795 0 1 2 +13 0 0 0.16368909908359128 0.84900594 0.14337672 0.007550044 0 1 2 +14 0 0 0.15933163163217298 0.8527135 0.13693798 0.009678988 0 1 2 +15 0 0 0.20822259663001871 0.81202626 0.16847935 0.018099552 0 1 2 +16 0 0 0.21723668717361938 0.8047395 0.17724717 0.016538456 0 1 2 +17 0 0 0.21707864066955088 0.8048667 0.18110135 0.014268022 0 1 2 +18 0 0 0.23144874426394701 0.79338336 0.188718 0.01717469 0 1 2 +19 0 0 0.21204888557965973 0.80892515 0.1755224 0.014853386 0 1 2 +20 0 0 0.22038504184482127 0.80220985 0.18447125 0.013839882 0 1 2 +21 0 0 0.2455567589196759 0.7822689 0.19803071 0.018726598 0 1 2 +22 0 0 0.15862707855033364 0.8533145 0.13755277 0.008518803 0 1 2 +23 0 0 0.32376478520521718 0.7234204 0.24993291 0.026990972 0 1 2 +24 0 0 0.23728116360449761 0.7887695 0.19633341 0.015370611 0 1 2 +25 0 0 0.22843370780417138 0.79577905 0.19070116 0.013262528 0 1 2 +26 0 0 0.27054019896153325 0.7629672 0.21736869 0.02023142 0 1 2 +27 0 0 0.19949086214486147 0.8191477 0.16933945 0.012093006 0 1 2 +28 0 0 0.19530796694222538 0.8225813 0.166539 0.011477919 0 1 2 +29 0 0 0.21853154983515735 0.8036981 0.18393002 0.013003357 0 1 2 +30 0 0 0.22305338845261874 0.80007213 0.18720335 0.013119485 0 1 2 +31 0 0 0.26104823770996771 0.77024376 0.2111881 0.019437576 0 1 2 +32 0 0 0.15411082498233339 0.857177 0.13288182 0.009111029 0 1 2 +33 0 0 0.16377952748963037 0.84892917 0.1392712 0.01077886 0 1 2 +34 0 0 0.18927891546323997 0.82755566 0.16317075 0.009776272 0 1 2 +35 0 0 0.18692560689818824 0.82950544 0.16045222 0.010186413 0 1 2 +36 0 0 0.18463589166135572 0.83140695 0.1582548 0.0107625425 0 1 2 +37 0 0 0.18927891546323997 0.82755566 0.16317075 0.009776272 0 1 2 +38 0 0 0.20160267074120711 0.81741965 0.17110164 0.010828887 0 1 2 +39 0 0 0.20307042096823624 0.81622076 0.17221838 0.012150835 0 1 2 +40 0 0 0.20844635205746181 0.8118446 0.17473942 0.013316462 0 1 2 +41 0 0 0.27330085883714583 0.76086384 0.22326645 0.014496124 0 1 2 +42 0 0 0.1933138149618753 0.8242233 0.16503218 0.010628331 0 1 2 +43 0 0 0.33868883712637471 0.7127042 0.2545966 0.0309349 0 1 2 +44 0 0 0.28186429019992376 0.75437605 0.2199256 0.02378716 0 1 2 +45 0 0 0.23855481852610427 0.7877655 0.19640626 0.014641503 0 1 2 +46 0 0 0.19478703925322305 0.8230099 0.16442285 0.012487361 0 1 2 +47 0 0 0.20173641138329881 0.81731033 0.17137061 0.011407288 0 1 2 +48 0 0 0.19170643499106257 0.8255492 0.16266319 0.011933463 0 1 2 +49 0 0 0.19866118495237869 0.8198276 0.16928534 0.01149064 0 1 2 +50 1 1 0.69296504535313552 0.5000911 0.39000574 0.11809637 1 2 0 +51 1 1 0.72211913094005886 0.48572186 0.4124787 0.111111 1 2 0 +52 1 1 0.74257566083556736 0.4758866 0.44323403 0.08918178 1 2 0 +53 1 1 0.58875267890508309 0.55501914 0.2788427 0.16242018 1 2 0 +54 1 1 0.6936646827577625 0.49974132 0.40895438 0.09816246 1 2 0 +55 1 1 0.6332083173602403 0.5308858 0.319683 0.14330727 1 2 0 +56 1 2 0.80040468364650408 0.4714638 0.44914716 0.08691386 2 1 0 +57 1 1 0.64443426699389328 0.52495944 0.31818488 0.14912984 1 0 2 +58 1 1 0.63792459802827473 0.5283879 0.3331588 0.13817137 1 2 0 +59 1 1 0.64394729286098484 0.52521515 0.30961776 0.15912524 1 2 0 +60 1 1 0.58012848318959576 0.55982643 0.2785041 0.1579764 1 0 2 +61 1 1 0.69392281850853987 0.49961233 0.3809702 0.12490331 1 2 0 +62 1 1 0.55792560747443609 0.5723952 0.2370053 0.18829401 1 0 2 +63 1 1 0.67006934840729138 0.5116731 0.3777513 0.11298229 1 2 0 +64 1 1 0.64396363500000775 0.52520657 0.25746092 0.2156204 1 2 0 +65 1 1 0.67383964679718267 0.50974756 0.36250323 0.13407421 1 2 0 +66 1 1 0.70520820338621382 0.4940057 0.4032462 0.108376935 1 2 0 +67 1 1 0.59230786888492815 0.55304945 0.2497409 0.19722393 1 0 2 +68 1 1 0.65005843181296841 0.5220153 0.38184637 0.0917979 1 2 0 +69 1 1 0.58432387959931353 0.55748266 0.23060052 0.21161926 1 0 2 +70 1 2 0.93076074140230303 0.5504721 0.39425367 0.056880735 2 1 0 +71 1 1 0.63132254698287349 0.5318879 0.28510487 0.17830437 1 2 0 +72 1 1 0.68980917574868994 0.5016718 0.42151535 0.07944365 1 2 0 +73 1 1 0.61242627367871161 0.54203415 0.2992107 0.15171584 1 2 0 +74 1 1 0.63705491357823907 0.52884763 0.31005612 0.15949327 1 2 0 +75 1 1 0.66724044054330434 0.5131226 0.35947827 0.1320263 1 2 0 +76 1 1 0.66516819266568061 0.51418704 0.38545328 0.105105296 1 2 0 +77 1 2 0.86786134305660678 0.5227722 0.4198485 0.058586687 2 1 0 +78 1 1 0.697282027618505 0.49793684 0.4020443 0.105632976 1 2 0 +79 1 1 0.63941292234221836 0.5276021 0.30470166 0.16353293 1 0 2 +80 1 1 0.58292306603094934 0.55826414 0.23556644 0.2051521 1 0 2 +81 1 1 0.60120184322687953 0.54815245 0.27547437 0.17247799 1 0 2 +82 1 1 0.60789895241224423 0.5444937 0.24415979 0.20928973 1 2 0 +83 1 2 0.77452094571976993 0.4817979 0.46092454 0.062509656 2 1 0 +84 1 1 0.7057304776402169 0.49374777 0.40238684 0.10868001 1 2 0 +85 1 2 0.79300614303245043 0.45822632 0.45248252 0.09833102 2 1 0 +86 1 1 0.72678760344330273 0.48345956 0.42667925 0.09861927 1 2 0 +87 1 1 0.59424353950347719 0.55197996 0.30586058 0.13495798 1 2 0 +88 1 1 0.64895935108537695 0.5225893 0.2943893 0.17855236 1 2 0 +89 1 1 0.60578735592724564 0.54564464 0.2812196 0.1685997 1 2 0 +90 1 1 0.59608992257788751 0.55096173 0.27534127 0.16702303 1 2 0 +91 1 1 0.67372798514007193 0.5098045 0.37284625 0.1206163 1 2 0 +92 1 1 0.59729060797438105 0.5503006 0.24914876 0.19754447 1 2 0 +93 1 1 0.63263790388930197 0.5311887 0.31309682 0.1488707 1 0 2 +94 1 1 0.62483482825269077 0.53534985 0.29635605 0.16058053 1 2 0 +95 1 1 0.63131761624848548 0.5318905 0.2671001 0.1972779 1 2 0 +96 1 1 0.64030851914936049 0.52712977 0.29998383 0.16769055 1 2 0 +97 1 1 0.6378120252633428 0.5284474 0.3092751 0.1597118 1 2 0 +98 1 1 0.66516656978472843 0.5141879 0.32090932 0.15677007 1 0 2 +99 1 1 0.63242225967493992 0.5313033 0.2911715 0.17120066 1 2 0 +100 2 2 0.19829896715055648 0.8201246 0.17368221 0.0062368084 2 1 0 +101 2 2 0.52498610781518373 0.5915636 0.37326652 0.035954762 2 1 0 +102 2 2 0.3407407532310523 0.7112433 0.27453977 0.015044473 2 1 0 +103 2 2 0.52057429965705759 0.5941792 0.3715722 0.033801537 2 1 0 +104 2 2 0.30970358459520808 0.7336644 0.25343123 0.013094116 2 1 0 +105 2 2 0.29094976644393783 0.7475532 0.2432034 0.009800421 2 1 0 +106 2 1 0.76728772489001817 0.4687222 0.4642706 0.07042906 1 2 0 +107 2 2 0.44188589454325078 0.642823 0.33604985 0.022054695 2 1 0 +108 2 2 0.5149521714355848 0.5975292 0.37382123 0.027675439 2 1 0 +109 2 2 0.18284764654039884 0.83289504 0.16138679 0.0060039973 2 1 0 +110 2 2 0.4459699668041292 0.640203 0.3286991 0.032119665 2 1 0 +111 2 2 0.50304188647519255 0.60468847 0.36408785 0.03160927 2 1 0 +112 2 2 0.37177305890065576 0.6895107 0.291309 0.019328838 2 1 0 +113 2 2 0.48953614760835817 0.6129106 0.35773036 0.030060844 2 1 0 +114 2 2 0.29316408877445871 0.74589974 0.24136522 0.013058035 2 1 0 +115 2 2 0.30009319908739385 0.7407492 0.24477336 0.014944793 2 1 0 +116 2 2 0.52479891734511053 0.5916743 0.37212884 0.036532886 2 1 0 +117 2 2 0.22283624718617615 0.8002459 0.19123352 0.008413073 2 1 0 +118 2 2 0.2241535254303231 0.7991924 0.19624045 0.004785996 2 1 0 +119 2 1 0.87274784250955761 0.50828177 0.41780192 0.07121508 1 2 0 +120 2 2 0.27441819275899937 0.7600142 0.2287033 0.011390187 2 1 0 +121 2 2 0.48522547863796006 0.6155584 0.3516373 0.033916466 2 1 0 +122 2 2 0.3254409055555631 0.72220886 0.26728806 0.0110192 2 1 0 +123 2 2 0.61063918499777836 0.5430037 0.40937084 0.04797497 2 1 0 +124 2 2 0.34419757992013356 0.7087889 0.2730288 0.01831223 2 1 0 +125 2 2 0.45983091871631876 0.6313904 0.34145266 0.027816722 2 1 0 +126 2 2 0.61811042048114551 0.5389619 0.40976956 0.05176377 2 1 0 +127 2 2 0.59452790008908529 0.551823 0.39778978 0.051540304 2 1 0 +128 2 2 0.37206320988495101 0.68931067 0.29314816 0.017891388 2 1 0 +129 2 2 0.60870189420969056 0.54405665 0.41200322 0.044889744 2 1 0 +130 2 2 0.41742602181118227 0.6587402 0.32220653 0.019624665 2 1 0 +131 2 2 0.30464783656769429 0.737383 0.24714842 0.015768483 2 1 0 +132 2 2 0.33115705253626715 0.7180924 0.2678029 0.014457943 2 1 0 +133 2 1 0.80647360566137283 0.482953 0.44642958 0.076360665 1 2 0 +134 2 1 0.81703446588011142 0.4920444 0.4417397 0.067763284 1 2 0 +135 2 2 0.25721579031812725 0.77320135 0.21887833 0.008458503 2 1 0 +136 2 2 0.24198383308520152 0.78506887 0.20457801 0.010242238 2 1 0 +137 2 2 0.52026367465991974 0.5943638 0.3687333 0.037257884 2 1 0 +138 2 2 0.60742780359224369 0.5447503 0.40198407 0.054496396 2 1 0 +139 2 2 0.37580665996697532 0.6867351 0.29258457 0.020977434 2 1 0 +140 2 2 0.25179778547921933 0.7774019 0.2131693 0.009634234 2 1 0 +141 2 2 0.31732326044803261 0.72809535 0.255903 0.016349927 2 1 0 +142 2 2 0.52498610781518373 0.5915636 0.37326652 0.035954762 2 1 0 +143 2 2 0.2625028787149627 0.76912415 0.22089085 0.010096038 2 1 0 +144 2 2 0.21257984430279722 0.80849576 0.18431029 0.007449273 2 1 0 +145 2 2 0.31396969120914769 0.73054117 0.25461537 0.015248882 2 1 0 +146 2 2 0.54647730899814428 0.5789858 0.38529754 0.036144704 2 1 0 +147 2 2 0.44535400270453684 0.64059746 0.33105242 0.028909232 2 1 0 +148 2 2 0.28674282658820827 0.75070477 0.23462024 0.014580021 2 1 0 +149 2 2 0.57039810426121063 0.56530035 0.38914245 0.04617037 2 1 0 diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Stacking-TrainTest-iris-out.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Stacking-TrainTest-iris-out.txt index 0ce2612e66..a8db8d83ca 100644 --- a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Stacking-TrainTest-iris-out.txt +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Stacking-TrainTest-iris-out.txt @@ -23,13 +23,13 @@ Beginning training model 4 of 5 Beginning optimization num vars: 15 improvement criterion: Mean Improvement -L1 regularization selected 12 of 15 weights. +L1 regularization selected 11 of 15 weights. Trainer 4 of 5 finished in %Time% Beginning training model 5 of 5 Beginning optimization num vars: 15 improvement criterion: Mean Improvement -L1 regularization selected 13 of 15 weights. +L1 regularization selected 11 of 15 weights. Trainer 5 of 5 finished in %Time% The number of instances used for stacking trainer is 43 Warning: The trainer specified for stacking wants normalization, but we do not currently allow this. @@ -50,15 +50,15 @@ TRUTH ||======================== Precision ||1.0000 |1.0000 |0.7692 | Accuracy(micro-avg): 0.900000 Accuracy(macro-avg): 0.900000 -Log-loss: 0.430568 -Log-loss reduction: 0.608080 +Log-loss: 0.431318 +Log-loss reduction: 0.607397 OVERALL RESULTS --------------------------------------- Accuracy(micro-avg): 0.900000 (0.0000) Accuracy(macro-avg): 0.900000 (0.0000) -Log-loss: 0.430568 (0.0000) -Log-loss reduction: 0.608080 (0.0000) +Log-loss: 0.431318 (0.0000) +Log-loss reduction: 0.607397 (0.0000) --------------------------------------- Physical memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Stacking-TrainTest-iris.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Stacking-TrainTest-iris.txt index 0124babe3c..cb0a108c7b 100644 --- a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Stacking-TrainTest-iris.txt +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/netcoreapp/WE-Stacking-TrainTest-iris.txt @@ -1,151 +1,151 @@ Instance Label Assigned Log-loss #1 Score #2 Score #3 Score #1 Class #2 Class #3 Class -0 0 0 0.18481972504974498 0.8312541 0.12271381 0.04603213 0 1 2 -1 0 0 0.20170184420340331 0.8173386 0.1337135 0.048947915 0 1 2 -2 0 0 0.18861319452864489 0.82810676 0.12531842 0.04657473 0 1 2 -3 0 0 0.20393994424857698 0.81551135 0.13505594 0.049432613 0 1 2 -4 0 0 0.18162509260959864 0.8339139 0.120627165 0.045458846 0 1 2 -5 0 0 0.23097517962067832 0.79375917 0.15060978 0.05563101 0 1 2 -6 0 0 0.20621873937701451 0.8136551 0.13616554 0.050179407 0 1 2 -7 0 0 0.19360066190337119 0.8239869 0.12833208 0.047681026 0 1 2 -8 0 0 0.20488191609285844 0.8147435 0.13579899 0.04945746 0 1 2 -9 0 0 0.18602063544070438 0.83025646 0.12377789 0.045965694 0 1 2 -10 0 0 0.18420339989786846 0.8317666 0.122187026 0.046046432 0 1 2 -11 0 0 0.19937852049807961 0.81923974 0.1319679 0.04879246 0 1 2 -12 0 0 0.18377711198429819 0.83212125 0.12240206 0.04547672 0 1 2 -13 0 0 0.1688368547093872 0.8446467 0.11278623 0.042566955 0 1 2 -14 0 0 0.16215089297001653 0.8503129 0.10794262 0.041744363 0 1 2 -15 0 0 0.19737341100351022 0.82088405 0.12967043 0.049445387 0 1 2 -16 0 0 0.2022455040673608 0.81689435 0.13315003 0.049955565 0 1 2 -17 0 0 0.2030596863285653 0.8162295 0.13410038 0.04967013 0 1 2 -18 0 0 0.2122805004503403 0.8087378 0.13949831 0.05176388 0 1 2 -19 0 0 0.1984388812130238 0.8200099 0.13097802 0.049012084 0 1 2 -20 0 0 0.20645237820388287 0.813465 0.13636237 0.050172564 0 1 2 -21 0 0 0.2237642133359905 0.7995036 0.14652169 0.053974565 0 1 2 -22 0 0 0.16270109788386145 0.8498452 0.108507775 0.0416471 0 1 2 -23 0 0 0.29319693221183341 0.74587524 0.18768671 0.06643816 0 1 2 -24 0 0 0.21951031248686975 0.8029119 0.14445785 0.052630328 0 1 2 -25 0 0 0.21506451477502536 0.8064894 0.14206003 0.05145059 0 1 2 -26 0 0 0.2457926089389037 0.7820844 0.16012962 0.057785966 0 1 2 -27 0 0 0.19043854084491313 0.82659656 0.12626418 0.047139246 0 1 2 -28 0 0 0.18812372674707292 0.8285122 0.124867655 0.046620015 0 1 2 -29 0 0 0.20671912581488569 0.81324804 0.1367144 0.050037563 0 1 2 -30 0 0 0.21075870510423816 0.8099695 0.13930826 0.050722137 0 1 2 -31 0 0 0.23788330856183545 0.7882947 0.15537657 0.056328777 0 1 2 -32 0 0 0.15887759465603593 0.8531008 0.10584394 0.041055378 0 1 2 -33 0 0 0.16544675611765486 0.847515 0.10996124 0.042523816 0 1 2 -34 0 0 0.18602063544070438 0.83025646 0.12377789 0.045965694 0 1 2 -35 0 0 0.1835048127175587 0.83234787 0.122060485 0.04559171 0 1 2 -36 0 0 0.17998171461201187 0.8352855 0.119628266 0.0450863 0 1 2 -37 0 0 0.18602063544070438 0.83025646 0.12377789 0.045965694 0 1 2 -38 0 0 0.19508003221461495 0.8227688 0.12955177 0.047679357 0 1 2 -39 0 0 0.19370910077330714 0.82389754 0.12839733 0.047705237 0 1 2 -40 0 0 0.1967334833132425 0.8214095 0.13015632 0.04843419 0 1 2 -41 0 0 0.247591012113099 0.78067917 0.16262333 0.0566975 0 1 2 -42 0 0 0.18831014528312648 0.82835776 0.12513448 0.04650773 0 1 2 -43 0 0 0.30498027579994319 0.7371379 0.1936697 0.06919234 0 1 2 -44 0 0 0.25256134263401325 0.77680856 0.16355811 0.059633296 0 1 2 -45 0 0 0.222810252987201 0.8002667 0.14677136 0.052961882 0 1 2 -46 0 0 0.1861524517475488 0.830147 0.12335987 0.046493188 0 1 2 -47 0 0 0.19420883832380909 0.8234859 0.12886404 0.047650035 0 1 2 -48 0 0 0.18410322381574865 0.83184993 0.122126475 0.046023536 0 1 2 -49 0 0 0.19122411097980588 0.82594746 0.12690006 0.047152534 0 1 2 -50 1 2 0.83683297642170662 0.45189413 0.43307993 0.11502599 2 1 0 -51 1 2 0.86783337594280763 0.47010145 0.41986024 0.11003827 2 1 0 -52 1 2 0.93428323925245738 0.51126915 0.39286736 0.09586359 2 1 0 -53 1 1 0.6480491393458786 0.5230652 0.32518908 0.15174574 1 2 0 -54 1 2 0.8554676532967751 0.47088045 0.42508435 0.10403518 2 1 0 -55 1 1 0.72061279164973846 0.48645407 0.37822276 0.13532309 1 2 0 -56 1 2 0.98743439174362635 0.53501433 0.37253124 0.09245437 2 1 0 -57 1 1 0.667659287305715 0.51290774 0.2862275 0.20086482 1 0 2 -58 1 1 0.74327479837518451 0.47555402 0.39353088 0.13091502 1 2 0 -59 1 1 0.69928553101847657 0.49694023 0.35608992 0.14696977 1 2 0 -60 1 1 0.61330208639307493 0.54155964 0.2469481 0.21149233 1 0 2 -61 1 1 0.80494976921103978 0.44711038 0.4323498 0.12053988 1 2 0 -62 1 1 0.60333297231991567 0.5469855 0.24247876 0.2105358 1 2 0 -63 1 1 0.80825593741383039 0.4456346 0.44061828 0.11374703 1 2 0 -64 1 1 0.67302112298475192 0.510165 0.30066988 0.18916516 1 2 0 -65 1 1 0.78642059354310501 0.4554722 0.41782764 0.12670006 1 2 0 -66 1 2 0.84646380288767165 0.46144047 0.42892903 0.109630495 2 1 0 -67 1 1 0.6375652667865348 0.5285778 0.25094622 0.22047597 1 2 0 -68 1 1 0.7937886503321171 0.4521286 0.44386703 0.104004346 1 2 0 -69 1 1 0.62678078877660259 0.5343091 0.26171502 0.20397584 1 2 0 -70 1 2 1.1803524715036871 0.620598 0.30717045 0.072231665 2 1 0 -71 1 1 0.68347946050470221 0.5048573 0.33492506 0.1602176 1 2 0 -72 1 2 0.88228443300498671 0.491925 0.41383645 0.094238445 2 1 0 -73 1 1 0.69890238562156548 0.49713066 0.3615451 0.14132433 1 2 0 -74 1 1 0.71302094137703598 0.4901612 0.36450952 0.14532934 1 2 0 -75 1 1 0.77878501817829404 0.4589633 0.41527632 0.12576029 1 2 0 -76 1 2 0.82148529235244627 0.45147872 0.43977797 0.10874335 2 1 0 -77 1 2 1.1207779859723854 0.5988077 0.32602605 0.075166285 2 1 0 -78 1 2 0.84253308990264597 0.46088678 0.43061835 0.10849492 2 1 0 -79 1 1 0.66611710150339865 0.51369935 0.27288014 0.21342051 1 0 2 -80 1 1 0.6211481141804488 0.5373272 0.25411865 0.20855416 1 2 0 -81 1 1 0.63286335903273561 0.531069 0.24458826 0.22434281 1 0 2 -82 1 1 0.64994208732468584 0.522076 0.29222375 0.18570015 1 2 0 -83 1 2 1.0167565026001655 0.55686325 0.36176643 0.0813704 2 1 0 -84 1 2 0.84479640446495841 0.4604885 0.42964482 0.109866776 2 1 0 -85 1 2 0.95379735615017791 0.51540005 0.38527521 0.099324845 2 1 0 -86 1 2 0.89646635613144265 0.4897553 0.40800887 0.10223584 2 1 0 -87 1 1 0.6818983213459886 0.5056562 0.36199653 0.13234733 1 2 0 -88 1 1 0.70079862028300921 0.49618888 0.34463507 0.15917611 1 2 0 -89 1 1 0.66098608824845284 0.5163419 0.32824793 0.15541014 1 2 0 -90 1 1 0.66119839829511173 0.5162323 0.3298869 0.15388076 1 2 0 -91 1 1 0.80126585356509961 0.44876054 0.43292412 0.11831529 1 2 0 -92 1 1 0.64518187605985866 0.5245671 0.2988823 0.17655061 1 2 0 -93 1 1 0.65791753821181764 0.5179288 0.28085864 0.20121251 1 0 2 -94 1 1 0.68765861584569643 0.5027518 0.34887746 0.14837083 1 2 0 -95 1 1 0.67876902213675616 0.507241 0.31932184 0.17343715 1 2 0 -96 1 1 0.70144984924003195 0.49586585 0.3525148 0.15161926 1 2 0 -97 1 1 0.71193179726649181 0.49069536 0.36367318 0.14563139 1 2 0 -98 1 1 0.68022047057949719 0.5065053 0.28840736 0.2050874 1 0 2 -99 1 1 0.68782022181335145 0.5026706 0.3420742 0.15525529 1 2 0 -100 2 2 0.16325307935556488 0.8493762 0.1190538 0.031570088 2 1 0 -101 2 2 0.40331268851515295 0.66810316 0.27097332 0.060923506 2 1 0 -102 2 2 0.24403950504641156 0.7834567 0.17442542 0.042117815 2 1 0 -103 2 2 0.38294958418802755 0.6818473 0.25942338 0.058729317 2 1 0 -104 2 2 0.22713015692038765 0.79681706 0.16318461 0.039998453 2 1 0 -105 2 2 0.20579717005041243 0.81399816 0.14892823 0.037073735 2 1 0 -106 2 2 0.63995466030369086 0.52731633 0.3853615 0.08732221 2 1 0 -107 2 2 0.30894847806844217 0.7342186 0.21610466 0.049676787 2 1 0 -108 2 2 0.37408830947761473 0.68791616 0.25556517 0.056518715 2 1 0 -109 2 2 0.15593196071841348 0.8556174 0.113773316 0.030609278 2 1 0 -110 2 2 0.3359117762615666 0.71468616 0.23098299 0.05433075 2 1 0 -111 2 2 0.37683848117154922 0.6860269 0.25623074 0.057742294 2 1 0 -112 2 2 0.27153323384777628 0.76220995 0.19208267 0.045707338 2 1 0 -113 2 2 0.37517759968763792 0.6871672 0.25557014 0.057262745 2 1 0 -114 2 2 0.22527077744714505 0.7983 0.1619161 0.03978388 2 1 0 -115 2 2 0.22755221208838697 0.79648083 0.16314988 0.040369198 2 1 0 -116 2 2 0.38945243435293203 0.6774277 0.2626792 0.059893034 2 1 0 -117 2 2 0.17289232753119629 0.8412282 0.12575305 0.03301885 2 1 0 -118 2 2 0.16843512380399916 0.8449861 0.12302486 0.03198899 2 1 0 -119 2 2 0.70979318130691527 0.4917459 0.41718784 0.09106621 2 1 0 -120 2 2 0.20613215521869691 0.81372553 0.14884703 0.03742757 2 1 0 -121 2 2 0.37539195633549916 0.68701994 0.25478923 0.058190938 2 1 0 -122 2 2 0.22543288714050602 0.7981706 0.1624201 0.039409187 2 1 0 -123 2 2 0.48212822762392815 0.6174679 0.3124032 0.07012886 2 1 0 -124 2 2 0.25081181497830374 0.7781688 0.17845584 0.043375447 2 1 0 -125 2 2 0.32661725177859197 0.7213598 0.22610879 0.05253134 2 1 0 -126 2 2 0.49302840299174844 0.6107739 0.317192 0.0720341 2 1 0 -127 2 2 0.46989319985744504 0.625069 0.30477363 0.07015733 2 1 0 -128 2 2 0.27102733111079919 0.76259565 0.19206518 0.04533926 2 1 0 -129 2 2 0.45260051814292879 0.63597214 0.29714867 0.06687913 2 1 0 -130 2 2 0.29329490955517062 0.74580216 0.20640816 0.047789603 2 1 0 -131 2 2 0.22065352582190517 0.8019945 0.15833211 0.039673436 2 1 0 -132 2 2 0.24324631515190609 0.78407836 0.17400321 0.04191848 2 1 0 -133 2 2 0.65317084227324096 0.5203931 0.38960916 0.089997634 2 1 0 -134 2 2 0.6492751089810922 0.52242434 0.39111754 0.0864581 2 1 0 -135 2 2 0.19067739274726869 0.82639915 0.13841815 0.03518274 2 1 0 -136 2 2 0.19018864398194518 0.82680315 0.13778706 0.035409875 2 1 0 -137 2 2 0.38616660720778828 0.6796573 0.26052547 0.059817314 2 1 0 -138 2 2 0.48527525060200399 0.61552775 0.31231016 0.07216221 2 1 0 -139 2 2 0.27584145520963421 0.75893325 0.1946325 0.046434242 2 1 0 -140 2 2 0.19410584561843153 0.8235707 0.14065166 0.03577768 2 1 0 -141 2 2 0.23975631070116138 0.7868196 0.17124212 0.04193833 2 1 0 -142 2 2 0.40331268851515295 0.66810316 0.27097332 0.060923506 2 1 0 -143 2 2 0.19746228997183493 0.8208111 0.14296111 0.036227815 2 1 0 -144 2 2 0.17192556278016694 0.84204185 0.12513997 0.03281808 2 1 0 -145 2 2 0.23660537300919193 0.7893027 0.16930264 0.04139455 2 1 0 -146 2 2 0.42092731290069557 0.6564378 0.2811992 0.062363073 2 1 0 -147 2 2 0.33280230232549579 0.7169119 0.22969246 0.05339575 2 1 0 -148 2 2 0.21879078303519184 0.8034898 0.1571611 0.039348997 2 1 0 -149 2 2 0.44329665369760785 0.64191675 0.2914011 0.066682145 2 1 0 +0 0 0 0.1841280161033135 0.8318293 0.12208592 0.046084676 0 1 2 +1 0 0 0.20090749301374131 0.8179881 0.13301137 0.04900055 0 1 2 +2 0 0 0.18785570758188253 0.8287343 0.12465183 0.04661387 0 1 2 +3 0 0 0.20319296468081996 0.81612074 0.13437924 0.0495 0 1 2 +4 0 0 0.18095537447172844 0.8344726 0.120015904 0.04551143 0 1 2 +5 0 0 0.23040006829196075 0.7942158 0.14996527 0.055818886 0 1 2 +6 0 0 0.2054693974875175 0.814265 0.13547595 0.050259054 0 1 2 +7 0 0 0.19290610738153935 0.8245594 0.1276929 0.047747646 0 1 2 +8 0 0 0.20407911477201698 0.81539786 0.13509177 0.049510293 0 1 2 +9 0 0 0.18531848236258605 0.83083963 0.12315122 0.046009023 0 1 2 +10 0 0 0.18355966767883491 0.8323022 0.12158485 0.046113 0 1 2 +11 0 0 0.19870880712461672 0.8197886 0.13133675 0.04887461 0 1 2 +12 0 0 0.18304682679585993 0.83272916 0.1217627 0.04550824 0 1 2 +13 0 0 0.16810702926017676 0.84526336 0.112166025 0.04257052 0 1 2 +14 0 0 0.16152645085958209 0.850844 0.10737925 0.041776653 0 1 2 +15 0 0 0.19680735605014343 0.82134885 0.12907156 0.049579524 0 1 2 +16 0 0 0.20152428691218111 0.8174837 0.13246553 0.050050817 0 1 2 +17 0 0 0.2023198581021248 0.8168336 0.13341635 0.04974993 0 1 2 +18 0 0 0.2116753786911878 0.80922735 0.13887097 0.051901795 0 1 2 +19 0 0 0.19778512358949829 0.82054615 0.1303455 0.049108278 0 1 2 +20 0 0 0.20579365527844828 0.814001 0.13572595 0.050273076 0 1 2 +21 0 0 0.22306657487824422 0.8000616 0.14582568 0.054112718 0 1 2 +22 0 0 0.16201498352516983 0.85042846 0.10791331 0.041658115 0 1 2 +23 0 0 0.29254266412791974 0.7463634 0.18692854 0.06670805 0 1 2 +24 0 0 0.21893359397537573 0.80337507 0.14385471 0.05277016 0 1 2 +25 0 0 0.21430800053476293 0.80709976 0.14136446 0.051535722 0 1 2 +26 0 0 0.24507593437861064 0.7826451 0.15940148 0.05795335 0 1 2 +27 0 0 0.18975972546948919 0.82715786 0.12563619 0.04720585 0 1 2 +28 0 0 0.18740895287201348 0.8291046 0.12422277 0.04667254 0 1 2 +29 0 0 0.20601203394587345 0.8138233 0.13605587 0.050120905 0 1 2 +30 0 0 0.21002838270963448 0.81056124 0.1386325 0.050806288 0 1 2 +31 0 0 0.23711085094434697 0.78890383 0.1546248 0.05647134 0 1 2 +32 0 0 0.15834904143746753 0.8535518 0.10534722 0.041101016 0 1 2 +33 0 0 0.16487978607214834 0.84799564 0.10942618 0.042578295 0 1 2 +34 0 0 0.18531848236258605 0.83083963 0.12315122 0.046009023 0 1 2 +35 0 0 0.18273164930087427 0.83299166 0.1213894 0.04561888 0 1 2 +36 0 0 0.17927023898131847 0.83588 0.1189935 0.045126434 0 1 2 +37 0 0 0.18531848236258605 0.83083963 0.12315122 0.046009023 0 1 2 +38 0 0 0.19428614410447417 0.82342225 0.12885937 0.047718238 0 1 2 +39 0 0 0.19301222983002841 0.8244719 0.12775627 0.047771946 0 1 2 +40 0 0 0.19597700103348209 0.82203114 0.12947127 0.048497595 0 1 2 +41 0 0 0.24654793182997081 0.7814939 0.16174 0.056766156 0 1 2 +42 0 0 0.18756007786920958 0.8289793 0.12447385 0.04654689 0 1 2 +43 0 0 0.3043427397471965 0.737608 0.19289002 0.06950207 0 1 2 +44 0 0 0.25209523882757284 0.7771707 0.1629459 0.059883352 0 1 2 +45 0 0 0.22196339055797881 0.8009447 0.14601132 0.05304393 0 1 2 +46 0 0 0.18555166536688333 0.8306459 0.12278006 0.046573967 0 1 2 +47 0 0 0.19346619667496159 0.8240977 0.12819992 0.047702335 0 1 2 +48 0 0 0.1834614178434604 0.832384 0.12152599 0.046089973 0 1 2 +49 0 0 0.19049312847557867 0.82655144 0.1262435 0.047204927 0 1 2 +50 1 2 0.83986604345277971 0.4540123 0.43176836 0.114219405 2 1 0 +51 1 2 0.87027072641413461 0.4717802 0.41883814 0.10938158 2 1 0 +52 1 2 0.93605663742294032 0.5124139 0.39217126 0.09541473 2 1 0 +53 1 1 0.64853024901904011 0.5228136 0.32637405 0.15081236 1 2 0 +54 1 2 0.85607947305037058 0.47157964 0.42482436 0.103596054 2 1 0 +55 1 1 0.72251078650909661 0.48553166 0.38013068 0.13433759 1 2 0 +56 1 2 0.98941121267603094 0.5361574 0.37179554 0.092047095 2 1 0 +57 1 1 0.667997862277938 0.5127341 0.28453752 0.20272839 1 0 2 +58 1 1 0.74556268813105619 0.47446725 0.39558536 0.12994748 1 2 0 +59 1 1 0.70047997966627007 0.496347 0.357602 0.14605103 1 2 0 +60 1 1 0.61342756390948472 0.5414917 0.24540576 0.21310258 1 0 2 +61 1 1 0.80663430279485715 0.44635785 0.4337911 0.11985106 1 2 0 +62 1 1 0.60389781088452088 0.54667664 0.24438684 0.20893656 1 2 0 +63 1 1 0.81002262594366148 0.444848 0.44209054 0.11306141 1 2 0 +64 1 1 0.67475586003244792 0.50928074 0.3029126 0.18780667 1 2 0 +65 1 1 0.78901216526200657 0.45429334 0.41989678 0.1258099 1 2 0 +66 1 2 0.84802173264119873 0.4626934 0.4282613 0.10904527 2 1 0 +67 1 1 0.63877991060776662 0.52793616 0.2534785 0.2185853 1 2 0 +68 1 1 0.79199695685176441 0.4529394 0.44331253 0.10374797 1 2 0 +69 1 1 0.62779946642433815 0.5337651 0.26381773 0.20241717 1 2 0 +70 1 2 1.1793181688534484 0.62028795 0.30748832 0.072223775 2 1 0 +71 1 1 0.68523836470678434 0.5039701 0.3369794 0.15905048 1 2 0 +72 1 2 0.88120472305637099 0.4917085 0.4142835 0.09400794 2 1 0 +73 1 1 0.70114181597529213 0.49601862 0.36381507 0.14016634 1 2 0 +74 1 1 0.71520206356962424 0.48909327 0.36667746 0.14422923 1 2 0 +75 1 1 0.78099918454788464 0.4579482 0.41713253 0.12491932 1 2 0 +76 1 2 0.82279845722905465 0.45263767 0.43920085 0.108161435 2 1 0 +77 1 2 1.1196272371232947 0.5984607 0.32640144 0.07513796 2 1 0 +78 1 2 0.84366085039850758 0.4619021 0.430133 0.10796483 2 1 0 +79 1 1 0.66672423981326034 0.51338756 0.27104852 0.21556388 1 0 2 +80 1 1 0.62196265757517133 0.5368897 0.25607866 0.20703162 1 2 0 +81 1 1 0.6334175055977973 0.5307748 0.24283952 0.22638574 1 0 2 +82 1 1 0.65142520814570792 0.5213023 0.29441264 0.18428507 1 2 0 +83 1 2 1.0151448014682309 0.5563498 0.36234996 0.08130031 2 1 0 +84 1 2 0.84635375139263191 0.4617478 0.42897624 0.109276086 2 1 0 +85 1 2 0.95640345671185811 0.51693135 0.38427246 0.098796226 2 1 0 +86 1 2 0.8983674344224869 0.49105656 0.40723395 0.10170938 2 1 0 +87 1 1 0.68226050149047401 0.5054731 0.36294943 0.13157739 1 2 0 +88 1 1 0.70310641817484731 0.4950451 0.34704182 0.15791313 1 2 0 +89 1 1 0.66196269503814287 0.5158379 0.32977995 0.15438214 1 2 0 +90 1 1 0.66271578141285847 0.5154496 0.33186498 0.15268537 1 2 0 +91 1 1 0.803457735189547 0.447778 0.43468067 0.117541365 1 2 0 +92 1 1 0.64652834530683412 0.5238613 0.30091637 0.17522238 1 2 0 +93 1 1 0.65820931543122041 0.5177777 0.2792198 0.20300257 1 0 2 +94 1 1 0.68919201891211956 0.5019815 0.35070387 0.14731465 1 2 0 +95 1 1 0.68111852319146848 0.50605065 0.32201785 0.17193149 1 2 0 +96 1 1 0.7035482121570098 0.49482644 0.35471654 0.15045694 1 2 0 +97 1 1 0.71410250814994625 0.48963135 0.36584362 0.144525 1 2 0 +98 1 1 0.68081774821669361 0.5062029 0.28690857 0.20688856 1 0 2 +99 1 1 0.68959996964588033 0.50177675 0.34411222 0.15411112 1 2 0 +100 2 2 0.16440678365436223 0.84839684 0.11974012 0.031862963 2 1 0 +101 2 2 0.40609917468075751 0.6662441 0.27254024 0.061215565 2 1 0 +102 2 2 0.2458655470066283 0.78202736 0.17552033 0.04245227 2 1 0 +103 2 2 0.38472685224582337 0.6806365 0.2604263 0.058937266 2 1 0 +104 2 2 0.2290053657821704 0.79532427 0.16432573 0.040349994 2 1 0 +105 2 2 0.20723032216584331 0.8128324 0.14978161 0.037386052 2 1 0 +106 2 2 0.64208411177086122 0.52619463 0.3864971 0.087308206 2 1 0 +107 2 2 0.31060960376714142 0.733 0.21705876 0.049941264 2 1 0 +108 2 2 0.37698654218512184 0.6859253 0.2572235 0.056851283 2 1 0 +109 2 2 0.15675327318419499 0.85491496 0.11422711 0.030858016 2 1 0 +110 2 2 0.33752728259911918 0.7135325 0.23190069 0.054566808 2 1 0 +111 2 2 0.37954021815835165 0.6841759 0.2577698 0.058054265 2 1 0 +112 2 2 0.27358957649361698 0.7606442 0.19330837 0.04604751 2 1 0 +113 2 2 0.37882923236196364 0.6846625 0.25766626 0.057671204 2 1 0 +114 2 2 0.22780084487132377 0.7962828 0.16349928 0.040218007 2 1 0 +115 2 2 0.22935316561817776 0.7950477 0.16424085 0.040711533 2 1 0 +116 2 2 0.390922276423114 0.6764327 0.2635034 0.06006389 2 1 0 +117 2 2 0.17341210812093677 0.84079105 0.12598889 0.033220116 2 1 0 +118 2 2 0.1699034872854441 0.84374624 0.12392624 0.03232753 2 1 0 +119 2 2 0.71178950557259157 0.49076518 0.41827542 0.090959296 2 1 0 +120 2 2 0.2076823544060864 0.8124651 0.14978218 0.03775265 2 1 0 +121 2 2 0.37819217830109125 0.6850988 0.2563862 0.05851496 2 1 0 +122 2 2 0.22715042885378808 0.7968009 0.16345328 0.03974589 2 1 0 +123 2 2 0.48446990757238495 0.61602366 0.31367502 0.07030122 2 1 0 +124 2 2 0.25222631815403124 0.77706885 0.17927824 0.043652885 2 1 0 +125 2 2 0.32760903367813859 0.7206447 0.22665083 0.052704424 2 1 0 +126 2 2 0.49496663552114289 0.60959125 0.3182506 0.0721581 2 1 0 +127 2 2 0.47115948268649155 0.624278 0.3054865 0.07023542 2 1 0 +128 2 2 0.27345684220388478 0.76074517 0.19353299 0.04572179 2 1 0 +129 2 2 0.45305048544352328 0.63568604 0.29742348 0.06689058 2 1 0 +130 2 2 0.29535913915082623 0.74426425 0.20762253 0.048113324 2 1 0 +131 2 2 0.22103501146332244 0.8016886 0.15847385 0.039837632 2 1 0 +132 2 2 0.24559996217679675 0.7822351 0.17544875 0.04231613 2 1 0 +133 2 2 0.65255813858606682 0.520712 0.3895665 0.089721434 2 1 0 +134 2 2 0.6490835661934814 0.5225244 0.39125416 0.08622134 2 1 0 +135 2 2 0.19220784522232714 0.82513535 0.1393483 0.035516433 2 1 0 +136 2 2 0.19151187493999464 0.8257098 0.13857752 0.03571268 2 1 0 +137 2 2 0.38734780852959166 0.67885494 0.2611855 0.05995947 2 1 0 +138 2 2 0.48648283435051609 0.6147849 0.31299442 0.07222071 2 1 0 +139 2 2 0.27773836068681401 0.757495 0.19575231 0.046752684 2 1 0 +140 2 2 0.19576718204038143 0.82220364 0.14167166 0.036124658 2 1 0 +141 2 2 0.24180960538509974 0.78520566 0.17248924 0.042305063 2 1 0 +142 2 2 0.40609917468075751 0.6662441 0.27254024 0.061215565 2 1 0 +143 2 2 0.19891262656195247 0.8196215 0.1438337 0.036544785 2 1 0 +144 2 2 0.17320611965414481 0.84096426 0.12591036 0.03312536 2 1 0 +145 2 2 0.23877812475853602 0.7875896 0.1706328 0.041777547 2 1 0 +146 2 2 0.42437484648962892 0.6541786 0.2831175 0.0627038 2 1 0 +147 2 2 0.3349058975761407 0.7154054 0.23090635 0.05368829 2 1 0 +148 2 2 0.22024848673187941 0.8023194 0.15802814 0.039652426 2 1 0 +149 2 2 0.44465947397136207 0.64104253 0.29216596 0.06679159 2 1 0 From de40d3cb7d122d1cae0f110ed436de0466649fd8 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Wed, 27 Mar 2024 13:08:07 -0700 Subject: [PATCH 17/26] Add net9.0 baselines --- ...on-CV-breast-cancer.PAVcalibration-out.txt | 58 ++ ...eptron-CV-breast-cancer.PAVcalibration.txt | 700 ++++++++++++++++++ ...ptron-CV-breast-cancer.calibrateRandom.txt | 700 ++++++++++++++++++ ...nTest-breast-cancer.PAVcalibration-out.txt | 39 + ...inTest-breast-cancer.PAVcalibration-rp.txt | 4 + ...TrainTest-breast-cancer.PAVcalibration.txt | 700 ++++++++++++++++++ ...Test-breast-cancer.calibrateRandom-out.txt | 38 + ...nTest-breast-cancer.calibrateRandom-rp.txt | 4 + ...rainTest-breast-cancer.calibrateRandom.txt | 700 ++++++++++++++++++ ...inTest-breast-cancer.nocalibration-out.txt | 39 + ...ainTest-breast-cancer.nocalibration-rp.txt | 4 + ...-TrainTest-breast-cancer.nocalibration.txt | 700 ++++++++++++++++++ .../net9.0/LDSVM-def-CV-breast-cancer-out.txt | 56 ++ .../net9.0/LDSVM-def-CV-breast-cancer.txt | 700 ++++++++++++++++++ .../net9.0/LDSVM-nob-CV-breast-cancer-out.txt | 56 ++ .../net9.0/LDSVM-nob-CV-breast-cancer-rp.txt | 4 + .../net9.0/LDSVM-nob-CV-breast-cancer.txt | 700 ++++++++++++++++++ .../LDSVM-nob-TrainTest-breast-cancer-out.txt | 38 + .../LDSVM-nob-TrainTest-breast-cancer-rp.txt | 4 + .../LDSVM-nob-TrainTest-breast-cancer.txt | 700 ++++++++++++++++++ ...gression-non-negative-CV-breast-cancer.txt | 700 ++++++++++++++++++ ...n-non-negative-TrainTest-breast-cancer.txt | 700 ++++++++++++++++++ .../MulticlassLogisticRegression-CV-iris.txt | 151 ++++ ...classLogisticRegression-TrainTest-iris.txt | 151 ++++ ...S-TrainTest-generatedRegressionDataset.txt | 501 +++++++++++++ .../WE-Average-TrainTest-breast-cancer.txt | 700 ++++++++++++++++++ .../WE-AvgPer-TrainTest-breast-cancer-out.txt | 50 ++ .../WE-AvgPer-TrainTest-breast-cancer-rp.txt | 4 + .../WE-AvgPer-TrainTest-breast-cancer.txt | 700 ++++++++++++++++++ .../net9.0/WE-Default-CV-breast-cancer.txt | 700 ++++++++++++++++++ .../WE-Hetero-TrainTest-breast-cancer-out.txt | 118 +++ .../WE-Hetero-TrainTest-breast-cancer-rp.txt | 4 + .../WE-Hetero-TrainTest-breast-cancer.txt | 700 ++++++++++++++++++ ...domFeature-TrainTest-breast-cancer-out.txt | 94 +++ ...ndomFeature-TrainTest-breast-cancer-rp.txt | 4 + .../net9.0/WE-Average-TrainTest-iris-out.txt | 61 ++ .../net9.0/WE-Average-TrainTest-iris-rp.txt | 4 + .../net9.0/WE-Average-TrainTest-iris.txt | 151 ++++ .../WE-Bootstrap-TrainTest-iris-out.txt | 151 ++++ .../net9.0/WE-Bootstrap-TrainTest-iris-rp.txt | 4 + .../net9.0/WE-Bootstrap-TrainTest-iris.txt | 151 ++++ .../net9.0/WE-Voting-TrainTest-iris-out.txt | 61 ++ 42 files changed, 11804 insertions(+) create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.PAVcalibration-out.txt create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.PAVcalibration.txt create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.calibrateRandom.txt create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-out.txt create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-rp.txt create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration.txt create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-out.txt create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-rp.txt create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom.txt create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-out.txt create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-rp.txt create mode 100644 test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration.txt create mode 100644 test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-def-CV-breast-cancer-out.txt create mode 100644 test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-def-CV-breast-cancer.txt create mode 100644 test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer-out.txt create mode 100644 test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer-rp.txt create mode 100644 test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer.txt create mode 100644 test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer-out.txt create mode 100644 test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer-rp.txt create mode 100644 test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer.txt create mode 100644 test/BaselineOutput/Common/LogisticRegression/net9.0/LogisticRegression-non-negative-CV-breast-cancer.txt create mode 100644 test/BaselineOutput/Common/LogisticRegression/net9.0/LogisticRegression-non-negative-TrainTest-breast-cancer.txt create mode 100644 test/BaselineOutput/Common/MulticlassLogisticRegression/net9.0/MulticlassLogisticRegression-CV-iris.txt create mode 100644 test/BaselineOutput/Common/MulticlassLogisticRegression/net9.0/MulticlassLogisticRegression-TrainTest-iris.txt create mode 100644 test/BaselineOutput/Common/OLS/net9.0/OLS-TrainTest-generatedRegressionDataset.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Average-TrainTest-breast-cancer.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer-out.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer-rp.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Default-CV-breast-cancer.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer-out.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer-rp.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-RandomFeature-TrainTest-breast-cancer-out.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-RandomFeature-TrainTest-breast-cancer-rp.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris-out.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris-rp.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris-out.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris-rp.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris.txt create mode 100644 test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Voting-TrainTest-iris-out.txt diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.PAVcalibration-out.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.PAVcalibration-out.txt new file mode 100644 index 0000000000..d6e8eb546b --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.PAVcalibration-out.txt @@ -0,0 +1,58 @@ +maml.exe CV tr=AveragedPerceptron threads=- cali=PAV dout=%Output% data=%Data% seed=1 +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Warning: Skipped 80 instances with missing features during training (over 10 iterations; 8 inst/iter) +Training calibrator. +PAV calibrator: piecewise function approximation has 5 components. +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Warning: Skipped 80 instances with missing features during training (over 10 iterations; 8 inst/iter) +Training calibrator. +PAV calibrator: piecewise function approximation has 6 components. +Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3785 (134.0/(134.0+220.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 133 | 1 | 0.9925 + negative || 9 | 211 | 0.9591 + ||====================== +Precision || 0.9366 | 0.9953 | +OVERALL 0/1 ACCURACY: 0.971751 +LOG LOSS/instance: Infinity +Test-set entropy (prior Log-Loss/instance): 0.956998 +LOG-LOSS REDUCTION (RIG): -Infinity +AUC: 0.994403 +Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3191 (105.0/(105.0+224.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 100 | 5 | 0.9524 + negative || 3 | 221 | 0.9866 + ||====================== +Precision || 0.9709 | 0.9779 | +OVERALL 0/1 ACCURACY: 0.975684 +LOG LOSS/instance: 0.227706 +Test-set entropy (prior Log-Loss/instance): 0.903454 +LOG-LOSS REDUCTION (RIG): 0.747961 +AUC: 0.997619 + +OVERALL RESULTS +--------------------------------------- +AUC: 0.996011 (0.0016) +Accuracy: 0.973718 (0.0020) +Positive precision: 0.953747 (0.0171) +Positive recall: 0.972459 (0.0201) +Negative precision: 0.986580 (0.0087) +Negative recall: 0.972849 (0.0138) +Log-loss: Infinity (NaN) +Log-loss reduction: -Infinity (NaN) +F1 Score: 0.962653 (0.0011) +AUPRC: 0.992269 (0.0025) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.PAVcalibration.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.PAVcalibration.txt new file mode 100644 index 0000000000..5556212934 --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.PAVcalibration.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +5 1 11.925824 1 -0 1 +6 0 -0.4527979 0.29372627 0.50170064939900283 0 +8 0 -3.7962546 1E-15 1.4415419267167138E-15 0 +9 0 -3.8130417 1E-15 1.4415419267167138E-15 0 +10 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +11 0 -4.606969 1E-15 1.4415419267167138E-15 0 +18 1 7.3475914 1 -0 1 +20 1 6.1389017 1 -0 1 +21 1 7.1486177 1 -0 1 +25 1 1.6632223 0.88235295 0.18057223417631088 1 +28 0 -4.606969 1E-15 1.4415419267167138E-15 0 +31 0 -4.2645645 1E-15 1.4415419267167138E-15 0 +32 1 7.198575 1 -0 1 +35 0 -4.606969 1E-15 1.4415419267167138E-15 0 +37 0 -1.714282 0.1 0.15200309583369792 0 +40 0 ? ? ? 0 +41 1 2.5451746 0.88235295 0.18057223417631088 1 +44 1 8.165841 1 -0 1 +45 0 -4.602255 1E-15 1.4415419267167138E-15 0 +46 1 5.6216097 1 -0 1 +48 0 -3.379683 1E-15 1.4415419267167138E-15 0 +50 1 2.8003244 0.88235295 0.18057223417631088 1 +51 1 0.14775372 0.6666667 0.58496245772549416 1 +52 1 4.696246 1 -0 1 +54 1 6.743868 1 -0 1 +56 1 6.5947094 1 -0 1 +60 1 2.2064123 0.88235295 0.18057223417631088 1 +63 1 0.8789625 0.6666667 0.58496245772549416 1 +64 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +66 0 -3.697434 1E-15 1.4415419267167138E-15 0 +68 1 9.899808 1 -0 1 +69 0 -4.3595524 1E-15 1.4415419267167138E-15 0 +70 0 -3.0557137 1E-15 1.4415419267167138E-15 0 +71 1 7.555621 1 -0 1 +72 0 -1.6769085 0.1 0.15200309583369792 0 +73 1 7.7111263 1 -0 1 +74 1 2.4994192 0.88235295 0.18057223417631088 1 +76 0 -3.8396955 1E-15 1.4415419267167138E-15 0 +77 0 -3.1438046 1E-15 1.4415419267167138E-15 0 +79 0 -4.4265766 1E-15 1.4415419267167138E-15 0 +82 0 -3.1870723 1E-15 1.4415419267167138E-15 0 +88 0 -3.697434 1E-15 1.4415419267167138E-15 0 +90 0 -4.54813 1E-15 1.4415419267167138E-15 0 +91 0 -4.5069323 1E-15 1.4415419267167138E-15 0 +92 0 -3.697434 1E-15 1.4415419267167138E-15 0 +93 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +95 0 -4.54813 1E-15 1.4415419267167138E-15 0 +96 0 -4.790498 1E-15 1.4415419267167138E-15 0 +97 0 -3.413869 1E-15 1.4415419267167138E-15 0 +98 1 9.294541 1 -0 1 +99 1 9.621308 1 -0 1 +100 1 5.1074314 1 -0 1 +102 0 -3.3471546 1E-15 1.4415419267167138E-15 0 +104 1 11.120679 1 -0 1 +105 1 2.1430407 0.88235295 0.18057223417631088 1 +106 1 8.747506 1 -0 1 +108 0 -4.513379 1E-15 1.4415419267167138E-15 0 +109 1 6.3912544 1 -0 1 +111 1 4.1208715 0.88235295 0.18057223417631088 1 +112 1 7.006653 1 -0 1 +113 1 9.855811 1 -0 1 +115 0 -3.4127908 1E-15 1.4415419267167138E-15 0 +117 1 8.638749 1 -0 1 +120 0 -4.0389752 1E-15 1.4415419267167138E-15 0 +121 0 -3.0156007 1E-15 1.4415419267167138E-15 0 +122 1 10.533509 1 -0 1 +123 1 4.6804914 1 -0 1 +125 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +128 1 4.9510326 1 -0 1 +129 0 -3.7898016 1E-15 1.4415419267167138E-15 0 +131 0 -4.2645645 1E-15 1.4415419267167138E-15 0 +132 1 9.206267 1 -0 1 +133 0 -4.1348257 1E-15 1.4415419267167138E-15 0 +137 0 -4.6658077 1E-15 1.4415419267167138E-15 0 +138 0 -3.6895585 1E-15 1.4415419267167138E-15 0 +141 0 -4.9493732 1E-15 1.4415419267167138E-15 0 +144 0 -4.606969 1E-15 1.4415419267167138E-15 0 +145 0 ? ? ? 0 +147 0 -4.2551055 1E-15 1.4415419267167138E-15 0 +150 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +151 1 5.0748987 1 -0 1 +152 1 8.881612 1 -0 1 +154 0 -5.232939 1E-15 1.4415419267167138E-15 0 +156 0 -4.3134584 1E-15 1.4415419267167138E-15 0 +161 0 -3.5088568 1E-15 1.4415419267167138E-15 0 +164 0 ? ? ? 0 +167 1 7.490014 1 -0 1 +169 0 -5.254455 1E-15 1.4415419267167138E-15 0 +171 0 -4.54813 1E-15 1.4415419267167138E-15 0 +173 1 14.494649 1 -0 1 +174 1 5.7514915 1 -0 1 +176 0 -4.2645645 1E-15 1.4415419267167138E-15 0 +177 1 6.138485 1 -0 1 +179 1 2.8749352 0.88235295 0.18057223417631088 1 +180 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +181 0 -5.232939 1E-15 1.4415419267167138E-15 0 +183 1 9.06396 1 -0 1 +187 1 12.004918 1 -0 1 +188 1 8.132482 1 -0 1 +189 0 -3.6207743 1E-15 1.4415419267167138E-15 0 +191 1 10.797475 1 -0 1 +192 0 -3.7562733 1E-15 1.4415419267167138E-15 0 +196 0 5.8725023 1 Infinity 1 +198 0 -5.232939 1E-15 1.4415419267167138E-15 0 +199 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +201 1 9.770608 1 -0 1 +202 0 -4.54813 1E-15 1.4415419267167138E-15 0 +204 0 -4.54813 1E-15 1.4415419267167138E-15 0 +205 1 12.087653 1 -0 1 +206 1 6.641532 1 -0 1 +207 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +209 0 -3.5942607 1E-15 1.4415419267167138E-15 0 +210 1 13.547517 1 -0 1 +211 1 9.089206 1 -0 1 +212 0 -4.54813 1E-15 1.4415419267167138E-15 0 +216 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +218 1 7.7499723 1 -0 1 +219 0 -2.4297438 1E-15 1.4415419267167138E-15 0 +223 1 5.4305964 1 -0 1 +226 1 9.166207 1 -0 1 +228 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +233 1 5.6998796 1 -0 1 +237 1 6.476473 1 -0 1 +239 1 5.1697493 1 -0 1 +240 0 -1.9057708 0.1 0.15200309583369792 0 +241 0 -3.8436403 1E-15 1.4415419267167138E-15 0 +242 0 -4.2645645 1E-15 1.4415419267167138E-15 0 +244 0 -4.54813 1E-15 1.4415419267167138E-15 0 +246 1 11.068807 1 -0 1 +247 1 3.1855068 0.88235295 0.18057223417631088 1 +248 0 -2.7545462 1E-15 1.4415419267167138E-15 0 +249 0 ? ? ? 0 +250 0 -4.655863 1E-15 1.4415419267167138E-15 0 +252 0 4.2659445 0.96023005 4.6521775310747371 1 +254 1 7.919628 1 -0 1 +257 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +258 0 -3.9809992 1E-15 1.4415419267167138E-15 0 +259 0 4.3407545 1 Infinity 1 +260 1 8.330071 1 -0 1 +262 1 9.629434 1 -0 1 +267 1 3.677929 0.88235295 0.18057223417631088 1 +268 1 8.609313 1 -0 1 +269 0 -4.54813 1E-15 1.4415419267167138E-15 0 +271 0 -3.413869 1E-15 1.4415419267167138E-15 0 +272 1 3.677929 0.88235295 0.18057223417631088 1 +275 0 ? ? ? 0 +276 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +277 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +278 0 -4.54813 1E-15 1.4415419267167138E-15 0 +279 1 6.636299 1 -0 1 +280 0 -3.9809992 1E-15 1.4415419267167138E-15 0 +283 1 5.5930414 1 -0 1 +284 1 6.3683043 1 -0 1 +285 1 13.342931 1 -0 1 +288 1 2.133195 0.88235295 0.18057223417631088 1 +290 0 -5.232939 1E-15 1.4415419267167138E-15 0 +291 0 -4.54813 1E-15 1.4415419267167138E-15 0 +293 1 4.659646 1 -0 1 +296 0 2.0631151 0.88235295 3.0874629272416674 1 +297 0 ? ? ? 0 +299 1 6.110528 1 -0 1 +300 1 6.6979446 1 -0 1 +301 0 -4.54813 1E-15 1.4415419267167138E-15 0 +303 0 -4.54813 1E-15 1.4415419267167138E-15 0 +304 1 4.865142 1 -0 1 +308 1 6.804652 1 -0 1 +309 0 -1.4760427 0.1 0.15200309583369792 0 +311 0 -5.232939 1E-15 1.4415419267167138E-15 0 +312 1 3.2489738 0.88235295 0.18057223417631088 1 +314 0 -5.0709267 1E-15 1.4415419267167138E-15 0 +316 1 4.409379 1 -0 1 +317 1 8.798547 1 -0 1 +319 0 2.7989082 0.88235295 3.0874629272416674 1 +321 0 ? ? ? 0 +323 1 4.8855343 1 -0 1 +327 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +328 1 4.1213064 0.88235295 0.18057223417631088 1 +329 1 6.918497 1 -0 1 +331 0 -3.1255894 1E-15 1.4415419267167138E-15 0 +332 0 -2.8317442 1E-15 1.4415419267167138E-15 0 +333 1 4.9879713 1 -0 1 +336 1 5.3119774 1 -0 1 +338 0 -5.0709267 1E-15 1.4415419267167138E-15 0 +343 0 -5.232939 1E-15 1.4415419267167138E-15 0 +344 1 9.373975 1 -0 1 +346 0 -2.8051786 1E-15 1.4415419267167138E-15 0 +347 0 -5.0361757 1E-15 1.4415419267167138E-15 0 +348 1 0.09843826 0.6666667 0.58496245772549416 1 +349 1 3.5435781 0.88235295 0.18057223417631088 1 +350 0 -3.7809262 1E-15 1.4415419267167138E-15 0 +352 0 1.5155859 0.88235295 3.0874629272416674 1 +353 1 9.391767 1 -0 1 +354 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +355 0 -3.6708689 1E-15 1.4415419267167138E-15 0 +358 1 6.2604895 1 -0 1 +360 1 15.035466 1 -0 1 +361 1 6.5571547 1 -0 1 +366 1 13.467437 1 -0 1 +368 0 -4.532379 1E-15 1.4415419267167138E-15 0 +370 0 -3.0555735 1E-15 1.4415419267167138E-15 0 +371 0 -4.532379 1E-15 1.4415419267167138E-15 0 +373 0 -3.5973978 1E-15 1.4415419267167138E-15 0 +376 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +377 0 -5.0709267 1E-15 1.4415419267167138E-15 0 +378 0 -3.4778075 1E-15 1.4415419267167138E-15 0 +379 0 -1.5486526 0.1 0.15200309583369792 0 +381 1 8.643491 1 -0 1 +383 0 -4.9493732 1E-15 1.4415419267167138E-15 0 +384 0 -4.9493732 1E-15 1.4415419267167138E-15 0 +387 0 -1.9204493 0.1 0.15200309583369792 0 +388 0 -4.418391 1E-15 1.4415419267167138E-15 0 +389 0 -2.9322662 1E-15 1.4415419267167138E-15 0 +391 1 9.3238 1 -0 1 +392 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +395 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +396 0 -3.9809992 1E-15 1.4415419267167138E-15 0 +398 0 -3.9080296 1E-15 1.4415419267167138E-15 0 +399 0 -4.305583 1E-15 1.4415419267167138E-15 0 +404 0 -4.499814 1E-15 1.4415419267167138E-15 0 +406 0 -3.4627619 1E-15 1.4415419267167138E-15 0 +409 0 -3.9731238 1E-15 1.4415419267167138E-15 0 +413 0 -3.063589 1E-15 1.4415419267167138E-15 0 +414 1 6.5001116 1 -0 1 +415 0 -0.5379734 0.15984297 0.25126909064706171 0 +416 1 8.981729 1 -0 1 +418 0 -1.7460232 0.1 0.15200309583369792 0 +419 0 -3.9546041 1E-15 1.4415419267167138E-15 0 +422 0 -2.1661267 1E-15 1.4415419267167138E-15 0 +423 0 -3.0557137 1E-15 1.4415419267167138E-15 0 +428 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +429 0 -4.606969 1E-15 1.4415419267167138E-15 0 +430 0 -4.248364 1E-15 1.4415419267167138E-15 0 +434 0 5.678664 1 Infinity 1 +436 1 4.8968487 1 -0 1 +439 0 -4.098677 1E-15 1.4415419267167138E-15 0 +440 1 7.6677313 1 -0 1 +441 0 -1.6798639 0.1 0.15200309583369792 0 +442 0 -4.0051136 1E-15 1.4415419267167138E-15 0 +449 1 10.087396 1 -0 1 +450 0 -3.9024794 1E-15 1.4415419267167138E-15 0 +451 0 -4.098677 1E-15 1.4415419267167138E-15 0 +452 0 -4.348057 1E-15 1.4415419267167138E-15 0 +453 1 8.068193 1 -0 1 +454 0 -4.3153887 1E-15 1.4415419267167138E-15 0 +455 1 0.4596901 0.6666667 0.58496245772549416 1 +456 1 8.942605 1 -0 1 +457 1 8.346686 1 -0 1 +464 0 -4.382242 1E-15 1.4415419267167138E-15 0 +465 1 8.953593 1 -0 1 +466 1 8.690221 1 -0 1 +467 1 7.4474792 1 -0 1 +474 0 -4.098677 1E-15 1.4415419267167138E-15 0 +480 0 -4.279069 1E-15 1.4415419267167138E-15 0 +482 1 14.295785 1 -0 1 +483 1 10.273198 1 -0 1 +484 0 -3.7483978 1E-15 1.4415419267167138E-15 0 +487 1 12.309562 1 -0 1 +489 1 -0.4385233 0.31616384 1.6612557298850197 0 +492 0 -4.22023 1E-15 1.4415419267167138E-15 0 +493 1 10.096398 1 -0 1 +495 0 -4.5037956 1E-15 1.4415419267167138E-15 0 +497 0 -4.155446 1E-15 1.4415419267167138E-15 0 +501 0 -4.0398383 1E-15 1.4415419267167138E-15 0 +502 0 -3.8809628 1E-15 1.4415419267167138E-15 0 +504 0 -5.232939 1E-15 1.4415419267167138E-15 0 +507 0 -3.8678274 1E-15 1.4415419267167138E-15 0 +510 0 -5.232939 1E-15 1.4415419267167138E-15 0 +513 0 -4.5037956 1E-15 1.4415419267167138E-15 0 +514 1 9.508458 1 -0 1 +517 0 -5.0709267 1E-15 1.4415419267167138E-15 0 +519 1 6.943595 1 -0 1 +520 0 -5.0525465 1E-15 1.4415419267167138E-15 0 +521 0 -4.2018504 1E-15 1.4415419267167138E-15 0 +522 1 4.3094034 0.9868348 0.019119467566511453 1 +523 1 6.7598047 1 -0 1 +527 0 -3.697434 1E-15 1.4415419267167138E-15 0 +528 0 -2.955677 1E-15 1.4415419267167138E-15 0 +529 0 -4.22023 1E-15 1.4415419267167138E-15 0 +531 0 -3.4627619 1E-15 1.4415419267167138E-15 0 +532 0 -4.7285223 1E-15 1.4415419267167138E-15 0 +533 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +534 0 -4.606969 1E-15 1.4415419267167138E-15 0 +535 0 -3.6886954 1E-15 1.4415419267167138E-15 0 +538 0 -4.0398383 1E-15 1.4415419267167138E-15 0 +539 0 -3.4727077 1E-15 1.4415419267167138E-15 0 +540 0 -3.3012362 1E-15 1.4415419267167138E-15 0 +541 0 -4.6658077 1E-15 1.4415419267167138E-15 0 +544 0 -3.7475343 1E-15 1.4415419267167138E-15 0 +546 1 11.016191 1 -0 1 +547 0 -5.1297655 1E-15 1.4415419267167138E-15 0 +548 0 -4.8462 1E-15 1.4415419267167138E-15 0 +549 1 5.5535192 1 -0 1 +557 0 -3.7809262 1E-15 1.4415419267167138E-15 0 +558 0 -4.606969 1E-15 1.4415419267167138E-15 0 +559 0 -3.7562733 1E-15 1.4415419267167138E-15 0 +560 0 -3.413869 1E-15 1.4415419267167138E-15 0 +561 0 -3.413869 1E-15 1.4415419267167138E-15 0 +563 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +565 1 11.259841 1 -0 1 +566 0 -3.6307197 1E-15 1.4415419267167138E-15 0 +569 1 9.242663 1 -0 1 +577 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +578 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +581 1 8.654241 1 -0 1 +582 1 8.128011 1 -0 1 +584 0 -3.123908 1E-15 1.4415419267167138E-15 0 +586 1 12.971318 1 -0 1 +590 1 4.1888065 0.9130077 0.13130110439503334 1 +593 0 -3.7483978 1E-15 1.4415419267167138E-15 0 +594 1 5.7850237 1 -0 1 +600 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +602 0 -4.0398383 1E-15 1.4415419267167138E-15 0 +604 1 5.138942 1 -0 1 +606 0 -4.223367 1E-15 1.4415419267167138E-15 0 +607 0 -5.232939 1E-15 1.4415419267167138E-15 0 +609 0 -4.098677 1E-15 1.4415419267167138E-15 0 +612 1 15.565835 1 -0 1 +613 0 -4.1533766 1E-15 1.4415419267167138E-15 0 +614 0 -4.787361 1E-15 1.4415419267167138E-15 0 +617 0 ? ? ? 0 +618 0 -4.0398383 1E-15 1.4415419267167138E-15 0 +619 0 -3.7562733 1E-15 1.4415419267167138E-15 0 +621 0 0.0042419434 0.6666667 1.5849625867124844 1 +622 0 -2.5211115 1E-15 1.4415419267167138E-15 0 +624 0 -3.7777896 1E-15 1.4415419267167138E-15 0 +627 0 -3.1292257 1E-15 1.4415419267167138E-15 0 +629 0 -4.382242 1E-15 1.4415419267167138E-15 0 +633 1 4.158327 0.89434874 0.16109059211947238 1 +634 0 -4.6658077 1E-15 1.4415419267167138E-15 0 +638 0 -4.382242 1E-15 1.4415419267167138E-15 0 +639 0 -3.7809262 1E-15 1.4415419267167138E-15 0 +641 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +642 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +644 0 -4.9493732 1E-15 1.4415419267167138E-15 0 +645 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +649 0 -4.3234034 1E-15 1.4415419267167138E-15 0 +652 0 -3.5676956 1E-15 1.4415419267167138E-15 0 +653 0 -4.0398383 1E-15 1.4415419267167138E-15 0 +654 0 -3.9809992 1E-15 1.4415419267167138E-15 0 +656 0 -3.7562733 1E-15 1.4415419267167138E-15 0 +657 0 0.6928797 0.6666667 1.5849625867124844 1 +660 0 -4.8905344 1E-15 1.4415419267167138E-15 0 +661 0 -3.697434 1E-15 1.4415419267167138E-15 0 +665 0 -5.232939 1E-15 1.4415419267167138E-15 0 +668 1 3.362691 0.88235295 0.18057223417631088 1 +670 1 6.9357576 1 -0 1 +678 0 -5.232939 1E-15 1.4415419267167138E-15 0 +679 0 -4.9493732 1E-15 1.4415419267167138E-15 0 +680 1 15.087735 1 -0 1 +681 1 9.870067 1 -0 1 +682 0 -3.2252913 1E-15 1.4415419267167138E-15 0 +683 0 -5.232939 1E-15 1.4415419267167138E-15 0 +685 0 -5.232939 1E-15 1.4415419267167138E-15 0 +688 0 -4.382242 1E-15 1.4415419267167138E-15 0 +689 0 -3.214981 1E-15 1.4415419267167138E-15 0 +691 1 5.504386 1 -0 1 +692 0 -4.6658077 1E-15 1.4415419267167138E-15 0 +693 0 -4.035124 1E-15 1.4415419267167138E-15 0 +694 0 -4.0303783 1E-15 1.4415419267167138E-15 0 +696 1 7.6071005 1 -0 1 +697 1 5.0033855 1 -0 1 +698 1 6.1210947 1 -0 1 +0 0 -3.4721127 1E-15 1.4415419267167138E-15 0 +1 0 2.4163914 0.9047619 3.3923175087700881 1 +2 0 -4.045404 1E-15 1.4415419267167138E-15 0 +3 0 2.9251795 0.9047619 3.3923175087700881 1 +4 0 -3.5088272 1E-15 1.4415419267167138E-15 0 +7 0 -4.6705537 1E-15 1.4415419267167138E-15 0 +12 1 -0.34343147 0.51152503 0.96712324524188775 0 +13 0 -4.618695 1E-15 1.4415419267167138E-15 0 +14 1 7.3602133 1 -0 1 +15 1 0.6494312 0.6 0.73696553683865695 1 +16 0 -4.220706 1E-15 1.4415419267167138E-15 0 +17 0 -3.955172 1E-15 1.4415419267167138E-15 0 +19 0 -2.9890532 1E-15 1.4415419267167138E-15 0 +22 0 -4.7037654 1E-15 1.4415419267167138E-15 0 +23 1 ? ? ? 0 +24 0 -5.4043503 1E-15 1.4415419267167138E-15 0 +26 0 -4.390918 1E-15 1.4415419267167138E-15 0 +27 0 -3.7376466 1E-15 1.4415419267167138E-15 0 +29 0 -5.4339433 1E-15 1.4415419267167138E-15 0 +30 0 -4.649441 1E-15 1.4415419267167138E-15 0 +33 0 -4.6981187 1E-15 1.4415419267167138E-15 0 +34 0 -4.453028 1E-15 1.4415419267167138E-15 0 +36 1 7.8327723 1 -0 1 +38 1 4.92861 0.93333334 0.09953566740867692 1 +39 1 1.079258 0.6 0.73696553683865695 1 +42 1 6.8985863 1 -0 1 +43 1 -0.49528742 0.5 1 0 +47 0 -5.669884 1E-15 1.4415419267167138E-15 0 +49 1 5.3024054 0.93333334 0.09953566740867692 1 +53 1 5.116104 0.93333334 0.09953566740867692 1 +55 1 4.4195347 0.93333334 0.09953566740867692 1 +57 1 0.5701313 0.6 0.73696553683865695 1 +58 1 1.1371021 0.6 0.73696553683865695 1 +59 1 1.6442318 0.9047619 0.14438990028345636 1 +61 0 -5.2770567 1E-15 1.4415419267167138E-15 0 +62 1 5.7670774 1 -0 1 +65 1 2.7867756 0.9047619 0.14438990028345636 1 +67 1 2.9058523 0.9047619 0.14438990028345636 1 +75 0 -4.2910423 1E-15 1.4415419267167138E-15 0 +78 0 -3.652576 1E-15 1.4415419267167138E-15 0 +80 0 -3.2955704 1E-15 1.4415419267167138E-15 0 +81 0 -3.9975338 1E-15 1.4415419267167138E-15 0 +83 0 -2.9832687 1E-15 1.4415419267167138E-15 0 +84 1 6.5824003 1 -0 1 +85 1 4.7604074 0.93333334 0.09953566740867692 1 +86 1 1.466999 0.9047619 0.14438990028345636 1 +87 1 5.223544 0.93333334 0.09953566740867692 1 +89 0 -5.017977 1E-15 1.4415419267167138E-15 0 +94 0 -4.921291 1E-15 1.4415419267167138E-15 0 +101 1 -0.85990334 0.4151543 1.2682804255962932 0 +103 1 0.36755466 0.6 0.73696553683865695 1 +107 1 4.617798 0.93333334 0.09953566740867692 1 +110 0 -3.1312823 1E-15 1.4415419267167138E-15 0 +114 0 -2.8847933 1E-15 1.4415419267167138E-15 0 +116 0 -0.40737772 0.5 1 0 +118 0 -5.2584124 1E-15 1.4415419267167138E-15 0 +119 0 -3.7868814 1E-15 1.4415419267167138E-15 0 +124 1 5.0384197 0.93333334 0.09953566740867692 1 +126 1 6.4765244 1 -0 1 +127 0 -4.4382315 1E-15 1.4415419267167138E-15 0 +130 0 -3.2361717 1E-15 1.4415419267167138E-15 0 +134 0 -4.75809 1E-15 1.4415419267167138E-15 0 +135 0 -2.670867 1E-15 1.4415419267167138E-15 0 +136 0 -4.220706 1E-15 1.4415419267167138E-15 0 +139 0 ? ? ? 0 +140 0 -4.9692993 1E-15 1.4415419267167138E-15 0 +142 1 3.7043686 0.93333334 0.09953566740867692 1 +143 0 -4.364107 1E-15 1.4415419267167138E-15 0 +146 1 0.3614874 0.6 0.73696553683865695 1 +148 0 -2.3573046 1E-15 1.4415419267167138E-15 0 +149 1 8.756336 1 -0 1 +153 0 -3.7005844 1E-15 1.4415419267167138E-15 0 +155 1 3.1619139 0.9047619 0.14438990028345636 1 +157 0 -4.921291 1E-15 1.4415419267167138E-15 0 +158 0 ? ? ? 0 +159 1 10.452139 1 -0 1 +160 1 7.997595 1 -0 1 +162 0 -4.4382315 1E-15 1.4415419267167138E-15 0 +163 0 -3.669045 1E-15 1.4415419267167138E-15 0 +165 0 -3.3581352 1E-15 1.4415419267167138E-15 0 +166 1 6.3355865 1 -0 1 +168 0 -4.4382315 1E-15 1.4415419267167138E-15 0 +170 0 -4.9692993 1E-15 1.4415419267167138E-15 0 +172 0 -5.669884 1E-15 1.4415419267167138E-15 0 +175 1 6.0381775 1 -0 1 +178 0 -3.955172 1E-15 1.4415419267167138E-15 0 +182 0 -2.9890532 1E-15 1.4415419267167138E-15 0 +184 1 5.409273 1 -0 1 +185 0 -4.8669662 1E-15 1.4415419267167138E-15 0 +186 1 3.9876003 0.93333334 0.09953566740867692 1 +190 1 10.521242 1 -0 1 +193 0 -5.4043503 1E-15 1.4415419267167138E-15 0 +194 0 -4.4382315 1E-15 1.4415419267167138E-15 0 +195 0 -3.955172 1E-15 1.4415419267167138E-15 0 +197 0 -2.8025956 1E-15 1.4415419267167138E-15 0 +200 1 8.673521 1 -0 1 +203 0 -3.4721127 1E-15 1.4415419267167138E-15 0 +208 0 -5.3500257 1E-15 1.4415419267167138E-15 0 +213 1 12.300528 1 -0 1 +214 1 11.895983 1 -0 1 +215 1 6.6002197 1 -0 1 +217 0 -5.4043503 1E-15 1.4415419267167138E-15 0 +220 0 -5.181178 1E-15 1.4415419267167138E-15 0 +221 1 7.9662914 1 -0 1 +222 1 -2.148725 1E-15 49.828921418077073 0 +224 1 8.4735565 1 -0 1 +225 0 -5.669884 1E-15 1.4415419267167138E-15 0 +227 1 6.748211 1 -0 1 +229 1 10.504805 1 -0 1 +230 1 4.829338 0.93333334 0.09953566740867692 1 +231 1 6.912092 1 -0 1 +232 0 1.0722923 0.6 1.3219281808786905 1 +234 0 -2.7037287 1E-15 1.4415419267167138E-15 0 +235 0 ? ? ? 0 +236 1 9.440506 1 -0 1 +238 1 10.690647 1 -0 1 +243 0 -3.3019714 1E-15 1.4415419267167138E-15 0 +245 0 -2.8561125 1E-15 1.4415419267167138E-15 0 +251 1 7.355525 1 -0 1 +253 1 6.8985863 1 -0 1 +255 1 3.745205 0.93333334 0.09953566740867692 1 +256 0 -4.9692993 1E-15 1.4415419267167138E-15 0 +261 1 9.009869 1 -0 1 +263 1 7.14557 1 -0 1 +264 1 4.0145664 0.93333334 0.09953566740867692 1 +265 0 -2.5156918 1E-15 1.4415419267167138E-15 0 +266 1 7.325534 1 -0 1 +270 1 5.5723915 1 -0 1 +273 1 0.037317276 0.6 0.73696553683865695 1 +274 0 -4.2340226 1E-15 1.4415419267167138E-15 0 +281 0 -4.6981187 1E-15 1.4415419267167138E-15 0 +282 1 2.860157 0.9047619 0.14438990028345636 1 +286 1 12.544172 1 -0 1 +287 0 -4.75809 1E-15 1.4415419267167138E-15 0 +289 1 6.6595707 1 -0 1 +292 1 ? ? ? 0 +294 0 ? ? ? 0 +295 1 5.621521 1 -0 1 +298 0 -2.4584541 1E-15 1.4415419267167138E-15 0 +302 1 12.725582 1 -0 1 +305 1 8.040865 1 -0 1 +306 0 -5.4043503 1E-15 1.4415419267167138E-15 0 +307 0 -5.4043503 1E-15 1.4415419267167138E-15 0 +310 0 -5.24115 1E-15 1.4415419267167138E-15 0 +313 0 -5.935418 1E-15 1.4415419267167138E-15 0 +315 0 ? ? ? 0 +318 0 -5.567325 1E-15 1.4415419267167138E-15 0 +320 1 5.5611877 1 -0 1 +322 0 -4.4382315 1E-15 1.4415419267167138E-15 0 +324 0 -5.4043503 1E-15 1.4415419267167138E-15 0 +325 0 -3.7860875 1E-15 1.4415419267167138E-15 0 +326 1 3.6223297 0.9066624 0.14136263119498629 1 +330 1 4.9927444 0.93333334 0.09953566740867692 1 +334 1 5.514736 1 -0 1 +335 0 -5.935418 1E-15 1.4415419267167138E-15 0 +337 0 -5.4043503 1E-15 1.4415419267167138E-15 0 +339 1 5.1161976 0.93333334 0.09953566740867692 1 +340 1 5.5803356 1 -0 1 +341 0 -5.4043503 1E-15 1.4415419267167138E-15 0 +342 0 -5.4523587 1E-15 1.4415419267167138E-15 0 +345 0 -5.935418 1E-15 1.4415419267167138E-15 0 +351 0 -4.921291 1E-15 1.4415419267167138E-15 0 +356 1 -0.91277075 0.25679386 1.9613174061683403 0 +357 1 9.874601 1 -0 1 +359 1 4.653248 0.93333334 0.09953566740867692 1 +362 0 -3.5174994 1E-15 1.4415419267167138E-15 0 +363 0 -1.6840982 1E-15 1.4415419267167138E-15 0 +364 0 -4.921291 1E-15 1.4415419267167138E-15 0 +365 0 -5.186825 1E-15 1.4415419267167138E-15 0 +367 1 9.040358 1 -0 1 +369 0 -5.1140847 1E-15 1.4415419267167138E-15 0 +372 0 -3.9699688 1E-15 1.4415419267167138E-15 0 +374 0 -4.453028 1E-15 1.4415419267167138E-15 0 +375 0 -5.935418 1E-15 1.4415419267167138E-15 0 +380 0 -5.935418 1E-15 1.4415419267167138E-15 0 +382 0 -3.5015903 1E-15 1.4415419267167138E-15 0 +385 0 -3.4273868 1E-15 1.4415419267167138E-15 0 +386 1 5.084236 0.93333334 0.09953566740867692 1 +390 0 -5.446712 1E-15 1.4415419267167138E-15 0 +393 0 -5.8810935 1E-15 1.4415419267167138E-15 0 +394 0 -4.909887 1E-15 1.4415419267167138E-15 0 +397 0 -4.48624 1E-15 1.4415419267167138E-15 0 +400 1 7.313369 1 -0 1 +401 0 -4.9692993 1E-15 1.4415419267167138E-15 0 +402 0 -2.7191267 1E-15 1.4415419267167138E-15 0 +403 0 -3.7908158 1E-15 1.4415419267167138E-15 0 +405 0 -5.669884 1E-15 1.4415419267167138E-15 0 +407 0 -5.669884 1E-15 1.4415419267167138E-15 0 +408 0 -3.537508 1E-15 1.4415419267167138E-15 0 +410 0 -5.669884 1E-15 1.4415419267167138E-15 0 +411 0 ? ? ? 0 +412 1 7.6394253 1 -0 1 +417 0 -5.669884 1E-15 1.4415419267167138E-15 0 +420 0 -2.696971 1E-15 1.4415419267167138E-15 0 +421 1 9.498289 1 -0 1 +424 0 -4.9692993 1E-15 1.4415419267167138E-15 0 +425 1 11.849485 1 -0 1 +426 0 -2.2324486 1E-15 1.4415419267167138E-15 0 +427 1 4.1596622 0.93333334 0.09953566740867692 1 +431 0 -2.9302087 1E-15 1.4415419267167138E-15 0 +432 0 -3.5334377 1E-15 1.4415419267167138E-15 0 +433 0 -4.013695 1E-15 1.4415419267167138E-15 0 +435 1 7.0216722 1 -0 1 +437 0 -4.48624 1E-15 1.4415419267167138E-15 0 +438 0 -3.5384207 1E-15 1.4415419267167138E-15 0 +443 0 -5.54259 1E-15 1.4415419267167138E-15 0 +444 0 -2.6761098 1E-15 1.4415419267167138E-15 0 +445 0 -5.4523587 1E-15 1.4415419267167138E-15 0 +446 0 -5.935418 1E-15 1.4415419267167138E-15 0 +447 0 -4.0031805 1E-15 1.4415419267167138E-15 0 +448 0 -5.8810935 1E-15 1.4415419267167138E-15 0 +458 0 -3.7524433 1E-15 1.4415419267167138E-15 0 +459 0 -3.5017056 1E-15 1.4415419267167138E-15 0 +460 0 -3.5568357 1E-15 1.4415419267167138E-15 0 +461 0 -3.3979883 1E-15 1.4415419267167138E-15 0 +462 0 -3.0737762 1E-15 1.4415419267167138E-15 0 +463 0 -4.282031 1E-15 1.4415419267167138E-15 0 +468 0 -4.48624 1E-15 1.4415419267167138E-15 0 +469 0 -5.419147 1E-15 1.4415419267167138E-15 0 +470 0 -4.649441 1E-15 1.4415419267167138E-15 0 +471 0 -3.0737762 1E-15 1.4415419267167138E-15 0 +472 0 -3.4657965 1E-15 1.4415419267167138E-15 0 +473 0 -4.48624 1E-15 1.4415419267167138E-15 0 +475 0 -4.9692993 1E-15 1.4415419267167138E-15 0 +476 0 -4.2355022 1E-15 1.4415419267167138E-15 0 +477 0 -4.48624 1E-15 1.4415419267167138E-15 0 +478 0 -3.744658 1E-15 1.4415419267167138E-15 0 +479 1 6.673234 1 -0 1 +481 0 -2.5658808 1E-15 1.4415419267167138E-15 0 +485 0 -4.6490927 1E-15 1.4415419267167138E-15 0 +486 0 -4.649441 1E-15 1.4415419267167138E-15 0 +488 1 0.95910263 0.6 0.73696553683865695 1 +490 0 -5.935418 1E-15 1.4415419267167138E-15 0 +491 1 5.556223 1 -0 1 +494 0 -0.015696526 0.59426165 1.3013784104855217 0 +496 0 -5.8810935 1E-15 1.4415419267167138E-15 0 +498 0 -4.220706 1E-15 1.4415419267167138E-15 0 +499 0 -4.220706 1E-15 1.4415419267167138E-15 0 +500 0 -2.9890532 1E-15 1.4415419267167138E-15 0 +503 0 -3.955172 1E-15 1.4415419267167138E-15 0 +505 0 -4.4522543 1E-15 1.4415419267167138E-15 0 +506 1 8.766859 1 -0 1 +508 0 -4.0031805 1E-15 1.4415419267167138E-15 0 +509 0 -5.4523587 1E-15 1.4415419267167138E-15 0 +511 0 -3.7376466 1E-15 1.4415419267167138E-15 0 +512 0 -4.0031805 1E-15 1.4415419267167138E-15 0 +515 1 7.3749876 1 -0 1 +516 0 -5.8810935 1E-15 1.4415419267167138E-15 0 +518 0 -4.49539 1E-15 1.4415419267167138E-15 0 +524 0 -4.7037654 1E-15 1.4415419267167138E-15 0 +525 0 -4.746127 1E-15 1.4415419267167138E-15 0 +526 0 -4.48624 1E-15 1.4415419267167138E-15 0 +530 1 5.325205 0.93333334 0.09953566740867692 1 +536 0 -3.4721127 1E-15 1.4415419267167138E-15 0 +537 0 -3.2213755 1E-15 1.4415419267167138E-15 0 +542 0 -3.4861355 1E-15 1.4415419267167138E-15 0 +543 0 -4.220706 1E-15 1.4415419267167138E-15 0 +545 0 -3.7376466 1E-15 1.4415419267167138E-15 0 +550 0 -4.7037654 1E-15 1.4415419267167138E-15 0 +551 0 -5.4043503 1E-15 1.4415419267167138E-15 0 +552 0 -3.2360563 1E-15 1.4415419267167138E-15 0 +553 0 -1.4225531 1E-15 1.4415419267167138E-15 0 +554 0 -4.9692993 1E-15 1.4415419267167138E-15 0 +555 0 -1.7431297 1E-15 1.4415419267167138E-15 0 +556 0 -2.9508896 1E-15 1.4415419267167138E-15 0 +562 0 -5.4043503 1E-15 1.4415419267167138E-15 0 +564 0 -3.7509632 1E-15 1.4415419267167138E-15 0 +567 0 -3.4350505 1E-15 1.4415419267167138E-15 0 +568 1 3.5748348 0.9047619 0.14438990028345636 1 +570 1 6.466877 1 -0 1 +571 1 9.048693 1 -0 1 +572 0 -4.7037654 1E-15 1.4415419267167138E-15 0 +573 0 -5.669884 1E-15 1.4415419267167138E-15 0 +574 1 5.533702 1 -0 1 +575 0 -3.2213755 1E-15 1.4415419267167138E-15 0 +576 0 -3.7376466 1E-15 1.4415419267167138E-15 0 +579 0 -5.4043503 1E-15 1.4415419267167138E-15 0 +580 0 -3.4869094 1E-15 1.4415419267167138E-15 0 +583 0 -4.9692993 1E-15 1.4415419267167138E-15 0 +585 0 -5.935418 1E-15 1.4415419267167138E-15 0 +587 0 -3.5334377 1E-15 1.4415419267167138E-15 0 +588 1 4.6442137 0.93333334 0.09953566740867692 1 +589 0 -4.0031805 1E-15 1.4415419267167138E-15 0 +591 1 4.243067 0.93333334 0.09953566740867692 1 +592 1 4.8517904 0.93333334 0.09953566740867692 1 +595 0 -3.7376466 1E-15 1.4415419267167138E-15 0 +596 0 -3.9699688 1E-15 1.4415419267167138E-15 0 +597 0 -2.9706378 1E-15 1.4415419267167138E-15 0 +598 0 -4.7037654 1E-15 1.4415419267167138E-15 0 +599 0 -2.9381208 1E-15 1.4415419267167138E-15 0 +601 0 -5.6155596 1E-15 1.4415419267167138E-15 0 +603 1 3.1762495 0.9047619 0.14438990028345636 1 +605 1 8.159748 1 -0 1 +608 1 8.079367 1 -0 1 +610 1 6.972575 1 -0 1 +611 1 5.494137 1 -0 1 +615 0 -3.7192311 1E-15 1.4415419267167138E-15 0 +616 0 -4.7037654 1E-15 1.4415419267167138E-15 0 +620 0 -4.7037654 1E-15 1.4415419267167138E-15 0 +623 0 -5.935418 1E-15 1.4415419267167138E-15 0 +625 0 -3.343666 1E-15 1.4415419267167138E-15 0 +626 1 3.8647957 0.93333334 0.09953566740867692 1 +628 0 -5.4523587 1E-15 1.4415419267167138E-15 0 +630 0 -2.7601237 1E-15 1.4415419267167138E-15 0 +631 0 -3.7376466 1E-15 1.4415419267167138E-15 0 +632 0 -5.935418 1E-15 1.4415419267167138E-15 0 +635 0 -4.2170873 1E-15 1.4415419267167138E-15 0 +636 1 8.162586 1 -0 1 +637 0 -2.370799 1E-15 1.4415419267167138E-15 0 +640 0 -4.039895 1E-15 1.4415419267167138E-15 0 +643 0 -5.935418 1E-15 1.4415419267167138E-15 0 +646 0 -5.426158 1E-15 1.4415419267167138E-15 0 +647 0 -5.4890733 1E-15 1.4415419267167138E-15 0 +648 1 8.579456 1 -0 1 +650 0 -3.6219683 1E-15 1.4415419267167138E-15 0 +651 0 -4.965017 1E-15 1.4415419267167138E-15 0 +655 0 -4.7037654 1E-15 1.4415419267167138E-15 0 +658 1 7.546403 1 -0 1 +659 0 -5.935418 1E-15 1.4415419267167138E-15 0 +662 0 -5.1684093 1E-15 1.4415419267167138E-15 0 +663 0 -5.1684093 1E-15 1.4415419267167138E-15 0 +664 0 -4.2574205 1E-15 1.4415419267167138E-15 0 +666 0 -3.0702114 1E-15 1.4415419267167138E-15 0 +667 0 -4.4382315 1E-15 1.4415419267167138E-15 0 +669 1 6.9817877 1 -0 1 +671 0 -3.9565368 1E-15 1.4415419267167138E-15 0 +672 0 -4.921291 1E-15 1.4415419267167138E-15 0 +673 0 -3.289723 1E-15 1.4415419267167138E-15 0 +674 0 -5.669884 1E-15 1.4415419267167138E-15 0 +675 0 -3.7340279 1E-15 1.4415419267167138E-15 0 +676 0 -5.419147 1E-15 1.4415419267167138E-15 0 +677 0 -4.0031805 1E-15 1.4415419267167138E-15 0 +684 0 -5.935418 1E-15 1.4415419267167138E-15 0 +686 0 -5.935418 1E-15 1.4415419267167138E-15 0 +687 0 -4.295347 1E-15 1.4415419267167138E-15 0 +690 0 -5.4890733 1E-15 1.4415419267167138E-15 0 +695 0 -5.4523587 1E-15 1.4415419267167138E-15 0 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.calibrateRandom.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.calibrateRandom.txt new file mode 100644 index 0000000000..75264f66de --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-CV-breast-cancer.calibrateRandom.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +5 1 11.925824 0.99988866 0.00016064073906035849 1 +6 0 -0.4527979 0.28799012 0.49003084056125074 0 +8 0 -3.7962546 0.026382558 0.038573081321561092 0 +9 0 -3.8130417 0.02603617 0.038059897891843629 0 +10 0 -4.7285223 0.012591722 0.018281355447793919 0 +11 0 -4.606969 0.013873976 0.02015606392328537 0 +18 1 7.3475914 0.99550986 0.006492485535140746 1 +20 1 6.1389017 0.9881576 0.017186984398807038 1 +21 1 7.1486177 0.9947303 0.0076226814052998944 1 +25 1 1.6632223 0.6911742 0.53287870995530984 1 +28 0 -4.606969 0.013873976 0.02015606392328537 0 +31 0 -4.2645645 0.01821836 0.026525908568759592 0 +32 1 7.198575 0.99493784 0.0073217047835288053 1 +35 0 -4.606969 0.013873976 0.02015606392328537 0 +37 0 -1.714282 0.12729812 0.19643918434677948 0 +40 0 ? ? ? 0 +41 1 2.5451746 0.82034165 0.28570322258699443 1 +44 1 8.165841 0.9976778 0.0033541180091675089 1 +45 0 -4.602255 0.013926216 0.020232493670021917 0 +46 1 5.6216097 0.9821182 0.026031444169057363 1 +48 0 -3.379683 0.036561195 0.053735061214071053 0 +50 1 2.8003244 0.8487653 0.2365623958486662 1 +51 1 0.14775372 0.39660627 1.3342206221239763 1 +52 1 4.696246 0.9629518 0.054464539198696107 1 +54 1 6.743868 0.9927051 0.01056288089528439 1 +56 1 6.5947094 0.9917779 0.011911021598241851 1 +60 1 2.2064123 0.7763946 0.36513800059720591 1 +63 1 0.8789625 0.5427823 0.88155440156950038 1 +64 0 -4.8905344 0.011062949 0.016049403073119514 0 +66 0 -3.697434 0.028514327 0.041735375802187391 0 +68 1 9.899808 0.99942744 0.00082626923994757486 1 +69 0 -4.3595524 0.016894378 0.024581670707795066 0 +70 0 -3.0557137 0.046994247 0.069443170930934078 0 +71 1 7.555621 0.9962023 0.0054893674226172791 1 +72 0 -1.6769085 0.1306929 0.20206216584072867 0 +73 1 7.7111263 0.99664944 0.0048419465200441776 1 +74 1 2.4994192 0.8148248 0.29543816942188406 1 +76 0 -3.8396955 0.025495261 0.037258893517405786 0 +77 0 -3.1438046 0.043905538 0.064774930929348148 0 +79 0 -4.4265766 0.016017534 0.023295486850883066 0 +82 0 -3.1870723 0.042460296 0.062595786294511416 0 +88 0 -3.697434 0.028514327 0.041735375802187391 0 +90 0 -4.54813 0.014540091 0.021130915135437803 0 +91 0 -4.5069323 0.015025148 0.021841204430750202 0 +92 0 -3.697434 0.028514327 0.041735375802187391 0 +93 0 -4.8905344 0.011062949 0.016049403073119514 0 +95 0 -4.54813 0.014540091 0.021130915135437803 0 +96 0 -4.790498 0.011983716 0.017393274985562833 0 +97 0 -3.413869 0.0356 0.052296442645105176 0 +98 1 9.294541 0.99906635 0.0013475973096922743 1 +99 1 9.621308 0.99928296 0.001034846704541113 1 +100 1 5.1074314 0.97314864 0.039267908030735391 1 +102 0 -3.3471546 0.037498932 0.055139953861279192 0 +104 1 11.120679 0.99978656 0.00030796780451509731 1 +105 1 2.1430407 0.76737404 0.38199813744949118 1 +106 1 8.747506 0.99854773 0.0020967014982888315 1 +108 0 -4.513379 0.014948206 0.021728511485699098 0 +109 1 6.3912544 0.99032205 0.014030327944048709 1 +111 1 4.1208715 0.9422745 0.085780676489496735 1 +112 1 7.006653 0.9940932 0.0085470080951718218 1 +113 1 9.855811 0.9994067 0.00085621167564539895 1 +115 0 -3.4127908 0.03562994 0.05234123234660417 0 +117 1 8.638749 0.99841446 0.0022892705964363013 1 +120 0 -4.0389752 0.02178405 0.031775107173957186 0 +121 0 -3.0156007 0.048468217 0.071676249166093706 0 +122 1 10.533509 0.9996569 0.00049505099616089013 1 +123 1 4.6804914 0.9624947 0.055149541356901512 1 +125 0 -4.8905344 0.011062949 0.016049403073119514 0 +128 1 4.9510326 0.9696393 0.044479919754725615 1 +129 0 -3.7898016 0.026516901 0.03877216337418192 0 +131 0 -4.2645645 0.01821836 0.026525908568759592 0 +132 1 9.206267 0.99899733 0.0014472717652274012 1 +133 0 -4.1348257 0.020192469 0.029429713697453865 0 +137 0 -4.6658077 0.013237967 0.019225888046709876 0 +138 0 -3.6895585 0.028691238 0.041998118606553773 0 +141 0 -4.9493732 0.0105544245 0.015307740921272838 0 +144 0 -4.606969 0.013873976 0.02015606392328537 0 +145 0 ? ? ? 0 +147 0 -4.2551055 0.018355653 0.026727668422101055 0 +150 0 -4.7285223 0.012591722 0.018281355447793919 0 +151 1 5.0748987 0.97245276 0.040299926989301821 1 +152 1 8.881612 0.99869674 0.0018814265868558253 1 +154 0 -5.232939 0.0084102405 0.012184722644637337 0 +156 0 -4.3134584 0.01752461 0.02550682412744406 0 +161 0 -3.5088568 0.03305537 0.048494813867653343 0 +164 0 ? ? ? 0 +167 1 7.490014 0.99599624 0.0057878035564347926 1 +169 0 -5.254455 0.008266402 0.011975463236807289 0 +171 0 -4.54813 0.014540091 0.021130915135437803 0 +173 1 14.494649 0.99998605 2.0122110477710732E-05 1 +174 1 5.7514915 0.983872 0.023457464452347754 1 +176 0 -4.2645645 0.01821836 0.026525908568759592 0 +177 1 6.138485 0.98815364 0.017192727854047185 1 +179 1 2.8749352 0.8563466 0.22373324768939593 1 +180 0 -4.7285223 0.012591722 0.018281355447793919 0 +181 0 -5.232939 0.0084102405 0.012184722644637337 0 +183 1 9.06396 0.9988752 0.0016236556168074772 1 +187 1 12.004918 0.9998956 0.00015066466904787792 1 +188 1 8.132482 0.99761444 0.0034457424608055095 1 +189 0 -3.6207743 0.03028231 0.044363291377780306 0 +191 1 10.797475 0.99972284 0.00039991508636870194 1 +192 0 -3.7562733 0.027225697 0.039822976735077079 0 +196 0 5.8725023 0.9853531 6.0932620413355663 1 +198 0 -5.232939 0.0084102405 0.012184722644637337 0 +199 0 -4.3234034 0.01738671 0.025304343386339974 0 +201 1 9.770608 0.99936444 0.00091721700916576874 1 +202 0 -4.54813 0.014540091 0.021130915135437803 0 +204 0 -4.54813 0.014540091 0.021130915135437803 0 +205 1 12.087653 0.9999023 0.00014094666717597959 1 +206 1 6.641532 0.99208087 0.01147037136803414 1 +207 0 -4.7285223 0.012591722 0.018281355447793919 0 +209 0 -3.5942607 0.03091816 0.045309587634713357 0 +210 1 13.547517 0.99997 4.3254285101545484E-05 1 +211 1 9.089206 0.9988979 0.0015908564017088566 1 +212 0 -4.54813 0.014540091 0.021130915135437803 0 +216 0 -4.8905344 0.011062949 0.016049403073119514 0 +218 1 7.7499723 0.99675274 0.0046924303126549046 1 +219 0 -2.4297438 0.07561278 0.11343078610029891 0 +223 1 5.4305964 0.97919416 0.030333134890891962 1 +226 1 9.166207 0.99896437 0.0014948734814852957 1 +228 0 -4.7285223 0.012591722 0.018281355447793919 0 +233 1 5.6998796 0.9831963 0.02444858173710833 1 +237 1 6.476473 0.99096054 0.01310048672532375 1 +239 1 5.1697493 0.97443426 0.037363242337555155 1 +240 0 -1.9057708 0.11106791 0.16985489199973464 0 +241 0 -3.8436403 0.02541614 0.037141763788010818 0 +242 0 -4.2645645 0.01821836 0.026525908568759592 0 +244 0 -4.54813 0.014540091 0.021130915135437803 0 +246 1 11.068807 0.9997774 0.00032121335660955822 1 +247 1 3.1855068 0.8845627 0.17696373353849706 1 +248 0 -2.7545462 0.05918352 0.088014762213689604 0 +249 0 ? ? ? 0 +250 0 -4.655863 0.013343409 0.019380057057366337 0 +252 0 4.2659445 0.9483331 4.2746154342457627 1 +254 1 7.919628 0.99716777 0.0040918470836742334 1 +257 0 -4.3234034 0.01738671 0.025304343386339974 0 +258 0 -3.9809992 0.022805601 0.033282500904753695 0 +259 0 4.3407545 0.95121753 4.3574934457139323 1 +260 1 8.330071 0.99796593 0.0029375285520939812 1 +262 1 9.629434 0.99928766 0.0010280485312489948 1 +267 1 3.677929 0.91941935 0.1212050683714937 1 +268 1 8.609313 0.99837637 0.002344307364667635 1 +269 0 -4.54813 0.014540091 0.021130915135437803 0 +271 0 -3.413869 0.0356 0.052296442645105176 0 +272 1 3.677929 0.91941935 0.1212050683714937 1 +275 0 ? ? ? 0 +276 0 -4.3234034 0.01738671 0.025304343386339974 0 +277 0 -4.8905344 0.011062949 0.016049403073119514 0 +278 0 -4.54813 0.014540091 0.021130915135437803 0 +279 1 6.636299 0.9920476 0.011518738356558554 1 +280 0 -3.9809992 0.022805601 0.033282500904753695 0 +283 1 5.5930414 0.98170805 0.026634049742924003 1 +284 1 6.3683043 0.9901426 0.01429180181912721 1 +285 1 13.342931 0.9999646 5.1079751554952729E-05 1 +288 1 2.133195 0.7659501 0.38467771879185741 1 +290 0 -5.232939 0.0084102405 0.012184722644637337 0 +291 0 -4.54813 0.014540091 0.021130915135437803 0 +293 1 4.659646 0.9618815 0.056068896729700848 1 +296 0 2.0631151 0.7556402 2.0329211816413606 1 +297 0 ? ? ? 0 +299 1 6.110528 0.98788613 0.017583336461512396 1 +300 1 6.6979446 0.9924313 0.010960882920314604 1 +301 0 -4.54813 0.014540091 0.021130915135437803 0 +303 0 -4.54813 0.014540091 0.021130915135437803 0 +304 1 4.865142 0.967527 0.047626214248213887 1 +308 1 6.804652 0.99305254 0.010058042410768688 1 +309 0 -1.4760427 0.15027468 0.23493153853978993 0 +311 0 -5.232939 0.0084102405 0.012184722644637337 0 +312 1 3.2489738 0.88969976 0.16860953636335851 1 +314 0 -5.0709267 0.009575989 0.013881804577639753 0 +316 1 4.409379 0.9537285 0.068349469891991879 1 +317 1 8.798547 0.9986064 0.0020119654595983538 1 +319 0 2.7989082 0.84861827 2.7237369855870539 1 +321 0 ? ? ? 0 +323 1 4.8855343 0.968041 0.046859938697557786 1 +327 0 -4.8905344 0.011062949 0.016049403073119514 0 +328 1 4.1213064 0.94229364 0.08575138254712171 1 +329 1 6.918497 0.99365956 0.0091764494650447715 1 +331 0 -3.1255894 0.044527903 0.065714352919487407 0 +332 0 -2.8317442 0.05580239 0.082839263730817278 0 +333 1 4.9879713 0.97050625 0.043190589776101178 1 +336 1 5.3119774 0.9771479 0.033351185116665787 1 +338 0 -5.0709267 0.009575989 0.013881804577639753 0 +343 0 -5.232939 0.0084102405 0.012184722644637337 0 +344 1 9.373975 0.99912435 0.0012638519898626412 1 +346 0 -2.8051786 0.056944888 0.084586010770688697 0 +347 0 -5.0361757 0.009846162 0.014275404034477916 0 +348 1 0.09843826 0.38710517 1.3692025306601752 1 +349 1 3.5435781 0.91099644 0.13448268326126653 1 +350 0 -3.7809262 0.026702764 0.039047636108828568 0 +352 0 1.5155859 0.66513264 1.5783383412583629 1 +353 1 9.391767 0.99913687 0.001245778098200614 1 +354 0 -4.8905344 0.011062949 0.016049403073119514 0 +355 0 -3.6708689 0.029115347 0.042628190289617063 0 +358 1 6.2604895 0.9892544 0.015586495937560322 1 +360 1 15.035466 0.999991 1.2984748573554836E-05 1 +361 1 6.5571547 0.99152654 0.012276699619017499 1 +366 1 13.467437 0.999968 4.6178080787551932E-05 1 +368 0 -4.532379 0.014723691 0.021399727175722864 0 +370 0 -3.0555735 0.046999324 0.069450857564014035 0 +371 0 -4.532379 0.014723691 0.021399727175722864 0 +373 0 -3.5973978 0.030842256 0.045196591009649775 0 +376 0 -4.8905344 0.011062949 0.016049403073119514 0 +377 0 -5.0709267 0.009575989 0.013881804577639753 0 +378 0 -3.4778075 0.033867203 0.04970659025721904 0 +379 0 -1.5486526 0.14293146 0.22251751547439058 0 +381 1 8.643491 0.99842054 0.0022804855789229006 1 +383 0 -4.9493732 0.0105544245 0.015307740921272838 0 +384 0 -4.9493732 0.0105544245 0.015307740921272838 0 +387 0 -1.9204493 0.10990162 0.16796329622532433 0 +388 0 -4.418391 0.016122172 0.023448912843152432 0 +389 0 -2.9322662 0.051671736 0.076541560075192722 0 +391 1 9.3238 0.99908817 0.0013160954165418136 1 +392 0 -4.3234034 0.01738671 0.025304343386339974 0 +395 0 -4.3234034 0.01738671 0.025304343386339974 0 +396 0 -3.9809992 0.022805601 0.033282500904753695 0 +398 0 -3.9080296 0.024158021 0.035280548736868804 0 +399 0 -4.305583 0.017634574 0.025668308897692095 0 +404 0 -4.499814 0.015110557 0.021966308329195765 0 +406 0 -3.4627619 0.034267485 0.050304443015876471 0 +409 0 -3.9731238 0.02294793 0.033492644855224019 0 +413 0 -3.063589 0.04670991 0.069012796103739935 0 +414 1 6.5001116 0.9911302 0.01285354412946392 1 +415 0 -0.5379734 0.27407813 0.46211381508363975 0 +416 1 8.981729 0.99879795 0.0017352301830455481 1 +418 0 -1.7460232 0.124474354 0.1917786561625977 0 +419 0 -3.9546041 0.023286052 0.033991995638970728 0 +422 0 -2.1661267 0.091923326 0.13911397780821783 0 +423 0 -3.0557137 0.046994247 0.069443170930934078 0 +428 0 -4.8905344 0.011062949 0.016049403073119514 0 +429 0 -4.606969 0.013873976 0.02015606392328537 0 +430 0 -4.248364 0.018454121 0.026872392407670757 0 +434 0 5.678664 0.9829106 5.8707522055318933 1 +436 1 4.8968487 0.96832275 0.046440099194999938 1 +439 0 -4.098677 0.020778872 0.030293408561140194 0 +440 1 7.6677313 0.99653023 0.0050145176664260493 1 +441 0 -1.6798639 0.13042167 0.20161210337860253 0 +442 0 -4.0051136 0.022375137 0.032647118008373256 0 +449 1 10.087396 0.99950796 0.00071003308680625738 1 +450 0 -3.9024794 0.024264032 0.035437284415460464 0 +451 0 -4.098677 0.020778872 0.030293408561140194 0 +452 0 -4.348057 0.017049437 0.024809236682836164 0 +453 1 8.068193 0.9974875 0.0036293536545799203 1 +454 0 -4.3153887 0.01749776 0.02546739731233694 0 +455 1 0.4596901 0.45824033 1.1258236575305995 1 +456 1 8.942605 0.9987594 0.0017909346040298939 1 +457 1 8.346686 0.99799305 0.0028983232844613941 1 +464 0 -4.382242 0.01659238 0.024138560366440985 0 +465 1 8.953593 0.99877036 0.0017750926332974581 1 +466 1 8.690221 0.998479 0.0021959971123143552 1 +467 1 7.4474792 0.99585676 0.0059898462794332257 1 +474 0 -4.098677 0.020778872 0.030293408561140194 0 +480 0 -4.279069 0.018009793 0.026219457722364156 0 +482 1 14.295785 0.9999836 2.3647808303057529E-05 1 +483 1 10.273198 0.99957657 0.00061101174577425621 1 +484 0 -3.7483978 0.027394839 0.040073847273859001 0 +487 1 12.309562 0.99991834 0.00011781292611185561 1 +489 1 -0.4385233 0.29036236 1.784073657091122 0 +492 0 -4.22023 0.018870676 0.027484782205950953 0 +493 1 10.096398 0.99951154 0.00070487107661414573 1 +495 0 -4.5037956 0.015062725 0.021896244607923124 0 +497 0 -4.155446 0.019865256 0.028947997269570321 0 +501 0 -4.0398383 0.021769185 0.031753182964832191 0 +502 0 -3.8809628 0.02467931 0.036051434067825343 0 +504 0 -5.232939 0.0084102405 0.012184722644637337 0 +507 0 -3.8678274 0.024936229 0.03643151796310639 0 +510 0 -5.232939 0.0084102405 0.012184722644637337 0 +513 0 -4.5037956 0.015062725 0.021896244607923124 0 +514 1 9.508458 0.99921453 0.0011336389645461593 1 +517 0 -5.0709267 0.009575989 0.013881804577639753 0 +519 1 6.943595 0.99378616 0.0089926501548524794 1 +520 0 -5.0525465 0.009717959 0.014088618955572644 0 +521 0 -4.2018504 0.019147776 0.027892299787110719 0 +522 1 4.3094034 0.95002776 0.073958419155255584 1 +523 1 6.7598047 0.99279785 0.010428101441132083 1 +527 0 -3.697434 0.028514327 0.041735375802187391 0 +528 0 -2.955677 0.050752092 0.075143180755238653 0 +529 0 -4.22023 0.018870676 0.027484782205950953 0 +531 0 -3.4627619 0.034267485 0.050304443015876471 0 +532 0 -4.7285223 0.012591722 0.018281355447793919 0 +533 0 -4.3234034 0.01738671 0.025304343386339974 0 +534 0 -4.606969 0.013873976 0.02015606392328537 0 +535 0 -3.6886954 0.02871069 0.042027010565217231 0 +538 0 -4.0398383 0.021769185 0.031753182964832191 0 +539 0 -3.4727077 0.03400237 0.049908447053141368 0 +540 0 -3.3012362 0.038862113 0.05718467663392135 0 +541 0 -4.6658077 0.013237967 0.019225888046709876 0 +544 0 -3.7475343 0.027413446 0.040101449095009192 0 +546 1 11.016191 0.9997677 0.00033514712058262439 1 +547 0 -5.1297655 0.009135183 0.01323984985731042 0 +548 0 -4.8462 0.011462053 0.016631747558922043 0 +549 1 5.5535192 0.9811253 0.027490706489269167 1 +557 0 -3.7809262 0.026702764 0.039047636108828568 0 +558 0 -4.606969 0.013873976 0.02015606392328537 0 +559 0 -3.7562733 0.027225697 0.039822976735077079 0 +560 0 -3.413869 0.0356 0.052296442645105176 0 +561 0 -3.413869 0.0356 0.052296442645105176 0 +563 0 -4.3234034 0.01738671 0.025304343386339974 0 +565 1 11.259841 0.99980927 0.00027519848718779308 1 +566 0 -3.6307197 0.03004708 0.044013371380514872 0 +569 1 9.242663 0.99902636 0.0014053525670487904 1 +577 0 -4.8905344 0.011062949 0.016049403073119514 0 +578 0 -4.8905344 0.011062949 0.016049403073119514 0 +581 1 8.654241 0.9984342 0.0022607625481964131 1 +582 1 8.128011 0.99760586 0.0034581548754549166 1 +584 0 -3.123908 0.04458577 0.065801733133594345 0 +586 1 12.971318 0.9999522 6.8966691413912035E-05 1 +590 1 4.1888065 0.9451905 0.081322982207098438 1 +593 0 -3.7483978 0.027394839 0.040073847273859001 0 +594 1 5.7850237 0.98429656 0.022835041928152281 1 +600 0 -4.3234034 0.01738671 0.025304343386339974 0 +602 0 -4.0398383 0.021769185 0.031753182964832191 0 +604 1 5.138942 0.9738064 0.038293140457828272 1 +606 0 -4.223367 0.018823775 0.027415817997774653 0 +607 0 -5.232939 0.0084102405 0.012184722644637337 0 +609 0 -4.098677 0.020778872 0.030293408561140194 0 +612 1 15.565835 0.99999416 8.427174504939984E-06 1 +613 0 -4.1533766 0.019897858 0.028995985926563699 0 +614 0 -4.787361 0.012013778 0.017437172228630178 0 +617 0 ? ? ? 0 +618 0 -4.0398383 0.021769185 0.031753182964832191 0 +619 0 -3.7562733 0.027225697 0.039822976735077079 0 +621 0 0.0042419434 0.36919922 0.66474364251979601 1 +622 0 -2.5211115 0.070608765 0.10564205571811729 0 +624 0 -3.7777896 0.02676875 0.039145448981732493 0 +627 0 -3.1292257 0.04440299 0.065525755923725629 0 +629 0 -4.382242 0.01659238 0.024138560366440985 0 +633 1 4.158327 0.9438998 0.083294361183546328 1 +634 0 -4.6658077 0.013237967 0.019225888046709876 0 +638 0 -4.382242 0.01659238 0.024138560366440985 0 +639 0 -3.7809262 0.026702764 0.039047636108828568 0 +641 0 -4.3234034 0.01738671 0.025304343386339974 0 +642 0 -4.3234034 0.01738671 0.025304343386339974 0 +644 0 -4.9493732 0.0105544245 0.015307740921272838 0 +645 0 -4.3234034 0.01738671 0.025304343386339974 0 +649 0 -4.3234034 0.01738671 0.025304343386339974 0 +652 0 -3.5676956 0.0315682 0.046277640684522559 0 +653 0 -4.0398383 0.021769185 0.031753182964832191 0 +654 0 -3.9809992 0.022805601 0.033282500904753695 0 +656 0 -3.7562733 0.027225697 0.039822976735077079 0 +657 0 0.6928797 0.5052755 1.0153027124530103 1 +660 0 -4.8905344 0.011062949 0.016049403073119514 0 +661 0 -3.697434 0.028514327 0.041735375802187391 0 +665 0 -5.232939 0.0084102405 0.012184722644637337 0 +668 1 3.362691 0.8984041 0.15456354881831241 1 +670 1 6.9357576 0.9937469 0.0090496738948646453 1 +678 0 -5.232939 0.0084102405 0.012184722644637337 0 +679 0 -4.9493732 0.0105544245 0.015307740921272838 0 +680 1 15.087735 0.99999136 1.2468796069497877E-05 1 +681 1 9.870067 0.9994135 0.00084640287826628446 1 +682 0 -3.2252913 0.041221607 0.060730698781433959 0 +683 0 -5.232939 0.0084102405 0.012184722644637337 0 +685 0 -5.232939 0.0084102405 0.012184722644637337 0 +688 0 -4.382242 0.01659238 0.024138560366440985 0 +689 0 -3.214981 0.041552313 0.061228404039596784 0 +691 1 5.504386 0.98037547 0.028593709808496726 1 +692 0 -4.6658077 0.013237967 0.019225888046709876 0 +693 0 -4.035124 0.021850502 0.031873115017123851 0 +694 0 -4.0303783 0.021932656 0.031994290662136743 0 +696 1 7.6071005 0.99635655 0.0052659907809486433 1 +697 1 5.0033855 0.9708609 0.042663488663692947 1 +698 1 6.1210947 0.98798794 0.017434669917962976 1 +0 0 -3.4721127 0.04114936 0.060621989293231218 0 +1 0 2.4163914 0.8102446 2.3977873037870139 1 +2 0 -4.045404 0.026690748 0.039029825304374488 0 +3 0 2.9251795 0.86401767 2.8785088509147978 1 +4 0 -3.5088272 0.040032472 0.058942489732172955 0 +7 0 -4.6705537 0.016548794 0.024074619674936022 0 +12 1 -0.34343147 0.33084044 1.5957925090833267 0 +13 0 -4.618695 0.017221209 0.025061370644087304 0 +14 1 7.3602133 0.9951006 0.0070856869219221227 1 +15 1 0.6494312 0.5178032 0.9495242357357877 1 +16 0 -4.220706 0.023354555 0.034093183590717085 0 +17 0 -3.955172 0.028584344 0.041839357341208416 0 +19 0 -2.9890532 0.05890275 0.087584283244111522 0 +22 0 -4.7037654 0.016131792 0.023463019884524334 0 +23 1 ? ? ? 0 +24 0 -5.4043503 0.009396285 0.013620063190565418 0 +26 0 -4.390918 0.020506335 0.029891933800352901 0 +27 0 -3.7376466 0.03370043 0.0494575738020104 0 +29 0 -5.4339433 0.009183523 0.0133102348468353 0 +30 0 -4.649441 0.016819376 0.024471611493406355 0 +33 0 -4.6981187 0.016201954 0.023565905766532006 0 +34 0 -4.453028 0.019554092 0.028490056372964728 0 +36 1 7.8327723 0.9966079 0.004902085221172487 1 +38 1 4.92861 0.96814406 0.046706359368045791 1 +39 1 1.079258 0.60037446 0.73606548851031739 1 +42 1 6.8985863 0.9929881 0.010151652400815765 1 +43 1 -0.49528742 0.30512363 1.7125341958491029 0 +47 0 -5.669884 0.007649496 0.011078316226716699 0 +49 1 5.3024054 0.9760178 0.035020675578791655 1 +53 1 5.116104 0.9723645 0.040430893694146056 1 +55 1 4.4195347 0.95331246 0.068978947156115675 1 +57 1 0.5701313 0.5023233 0.99331181510286393 1 +58 1 1.1371021 0.6111654 0.71036521451913393 1 +59 1 1.6442318 0.7002284 0.51410253308935716 1 +61 0 -5.2770567 0.010368525 0.015036709335280767 0 +62 1 5.7670774 0.98319566 0.024449543808342401 1 +65 1 2.7867756 0.85080725 0.23309576864145995 1 +67 1 2.9058523 0.86223394 0.2138487480166619 1 +75 0 -4.2910423 0.022133546 0.032290643242935589 0 +78 0 -3.652576 0.03593296 0.052794624529770413 0 +80 0 -3.2955704 0.046948787 0.069374353985611328 0 +81 0 -3.9975338 0.027679633 0.040496353867953659 0 +83 0 -2.9832687 0.05915375 0.087969113977253724 0 +84 1 6.5824003 0.99104106 0.012983257475290413 1 +85 1 4.7604074 0.9638325 0.053145649556241237 1 +86 1 1.466999 0.67038584 0.57693642363214981 1 +87 1 5.223544 0.97453266 0.037217553175852479 1 +89 0 -5.017977 0.012665042 0.018388485721842051 0 +94 0 -4.921291 0.013645152 0.019821334809691313 0 +101 1 -0.85990334 0.2482707 2.0100140538749183 0 +103 1 0.36755466 0.46282858 1.1114501509356363 1 +107 1 4.617798 0.9597418 0.059281809358577985 1 +110 0 -3.1312823 0.053037077 0.078620155068192865 0 +114 0 -2.8847933 0.06358327 0.094777384611786961 0 +116 0 -0.40737772 0.31987607 0.55613045147256235 0 +118 0 -5.2584124 0.010519048 0.015256160233906227 0 +119 0 -3.7868814 0.032470126 0.047621887034091648 0 +124 1 5.0384197 0.9706861 0.042923294438382942 1 +126 1 6.4765244 0.9902761 0.014097276720975665 1 +127 0 -4.4382315 0.019776938 0.028818005735429076 0 +130 0 -3.2361717 0.049069244 0.072587803444637264 0 +134 0 -4.75809 0.015471875 0.022495674338028411 0 +135 0 -2.670867 0.074289635 0.11136721898934138 0 +136 0 -4.220706 0.023354555 0.034093183590717085 0 +139 0 ? ? ? 0 +140 0 -4.9692993 0.01314948 0.019096521726067024 0 +142 1 3.7043686 0.92112815 0.11852620676711542 1 +143 0 -4.364107 0.020931277 0.030517966172014473 0 +146 1 0.3614874 0.4616504 1.1151273539763837 1 +148 0 -2.3573046 0.09299236 0.14081338850031172 0 +149 1 8.756336 0.9983485 0.0023846173156615046 1 +153 0 -3.7005844 0.034656115 0.050885128687663553 0 +155 1 3.1619139 0.8843215 0.17735711249275102 1 +157 0 -4.921291 0.013645152 0.019821334809691313 0 +158 0 ? ? ? 0 +159 1 10.452139 0.99956036 0.00063441148415971371 1 +160 1 7.997595 0.9970165 0.0043107295502774808 1 +162 0 -4.4382315 0.019776938 0.028818005735429076 0 +163 0 -3.669045 0.035489924 0.052131785430299962 0 +165 0 -3.3581352 0.044809643 0.066139823021876529 0 +166 1 6.3355865 0.9891566 0.015729147554240097 1 +168 0 -4.4382315 0.019776938 0.028818005735429076 0 +170 0 -4.9692993 0.01314948 0.019096521726067024 0 +172 0 -5.669884 0.007649496 0.011078316226716699 0 +175 1 6.0381775 0.98635924 0.019814913404171304 1 +178 0 -3.955172 0.028584344 0.041839357341208416 0 +182 0 -2.9890532 0.05890275 0.087584283244111522 0 +184 1 5.409273 0.9778962 0.032246738266973496 1 +185 0 -4.8669662 0.014228281 0.020674502512045188 0 +186 1 3.9876003 0.93577826 0.095761382134015888 1 +190 1 10.521242 0.9995835 0.00060103256102277266 1 +193 0 -5.4043503 0.009396285 0.013620063190565418 0 +194 0 -4.4382315 0.019776938 0.028818005735429076 0 +195 0 -3.955172 0.028584344 0.041839357341208416 0 +197 0 -2.8025956 0.06751539 0.10084817709446191 0 +200 1 8.673521 0.99823827 0.002543887091121199 1 +203 0 -3.4721127 0.04114936 0.060621989293231218 0 +208 0 -5.3500257 0.009799641 0.014207621988617358 0 +213 1 12.300528 0.9998962 0.00014971866598967564 1 +214 1 11.895983 0.99985766 0.00020536190066155745 1 +215 1 6.6002197 0.99116385 0.012804525065380799 1 +217 0 -5.4043503 0.009396285 0.013620063190565418 0 +220 0 -5.181178 0.011165955 0.016199679762803301 0 +221 1 7.9662914 0.9969428 0.0044173368194155311 1 +222 1 -2.148725 0.107677184 3.2152155084388743 0 +224 1 8.4735565 0.997941 0.0029735466380567798 1 +225 0 -5.669884 0.007649496 0.011078316226716699 0 +227 1 6.748211 0.992121 0.011412038429712602 1 +229 1 10.504805 0.9995781 0.00060877502594878564 1 +230 1 4.829338 0.9656635 0.050407558021295686 1 +231 1 6.912092 0.9930612 0.010045486491117618 1 +232 0 1.0722923 0.59906816 1.3185711185935187 1 +234 0 -2.7037287 0.07254335 0.10864824861743642 0 +235 0 ? ? ? 0 +236 1 9.440506 0.9990316 0.0013977779753363091 1 +238 1 10.690647 0.9996351 0.00052653496634314379 1 +243 0 -3.3019714 0.046725545 0.069036458139849044 0 +245 0 -2.8561125 0.064930424 0.096854378999610469 0 +251 1 7.355525 0.99508274 0.0071116115660869242 1 +253 1 6.8985863 0.9929881 0.010151652400815765 1 +255 1 3.745205 0.9234149 0.11494910520956844 1 +256 0 -4.9692993 0.01314948 0.019096521726067024 0 +261 1 9.009869 0.9986448 0.0019564247198453513 1 +263 1 7.14557 0.9942114 0.0083754842729617477 1 +264 1 4.0145664 0.9370327 0.093828700465107676 1 +265 0 -2.5156918 0.083068565 0.12511423610705438 0 +266 1 7.325534 0.99496675 0.0072797874039680641 1 +270 1 5.5723915 0.9804892 0.028426363787945077 1 +273 1 0.037317276 0.39964524 1.3232081920564036 1 +274 0 -4.2340226 0.023118444 0.033744444283338601 0 +281 0 -4.6981187 0.016201954 0.023565905766532006 0 +282 1 2.860157 0.8579385 0.22105381364130766 1 +286 1 12.544172 0.9999142 0.00012374682432127928 1 +287 0 -4.75809 0.015471875 0.022495674338028411 0 +289 1 6.6595707 0.9915608 0.012226832918130892 1 +292 1 ? ? ? 0 +294 0 ? ? ? 0 +295 1 5.621521 0.98121005 0.027366079814265764 1 +298 0 -2.4584541 0.08653848 0.1305841369653 0 +302 1 12.725582 0.99992555 0.0001074071635410925 1 +305 1 8.040865 0.9971154 0.0041676501364772073 1 +306 0 -5.4043503 0.009396285 0.013620063190565418 0 +307 0 -5.4043503 0.009396285 0.013620063190565418 0 +310 0 -5.24115 0.010660341 0.015462184404232644 0 +313 0 -5.935418 0.0062253983 0.0090094240348259531 0 +315 0 ? ? ? 0 +318 0 -5.567325 0.008282307 0.011998599579238158 0 +320 1 5.5611877 0.98032105 0.028673793676042177 1 +322 0 -4.4382315 0.019776938 0.028818005735429076 0 +324 0 -5.4043503 0.009396285 0.013620063190565418 0 +325 0 -3.7860875 0.032489616 0.04765095016422824 0 +326 1 3.6223297 0.9163445 0.12603797604758099 1 +330 1 4.9927444 0.96965355 0.044458724475427286 1 +334 1 5.514736 0.97960865 0.029722574521978982 1 +335 0 -5.935418 0.0062253983 0.0090094240348259531 0 +337 0 -5.4043503 0.009396285 0.013620063190565418 0 +339 1 5.1161976 0.97236645 0.040427975332860647 1 +340 1 5.5803356 0.98060757 0.028252197193330195 1 +341 0 -5.4043503 0.009396285 0.013620063190565418 0 +342 0 -5.4523587 0.00905354 0.013120983569838577 0 +345 0 -5.935418 0.0062253983 0.0090094240348259531 0 +351 0 -4.921291 0.013645152 0.019821334809691313 0 +356 1 -0.91277075 0.24064313 2.0550328652513912 0 +357 1 9.874601 0.9993099 0.0009959512596500891 1 +359 1 4.653248 0.96079826 0.057694551223446147 1 +362 0 -3.5174994 0.039772928 0.058552483579195597 0 +363 0 -1.6840982 0.14783019 0.23078715020667279 0 +364 0 -4.921291 0.013645152 0.019821334809691313 0 +365 0 -5.186825 0.011117355 0.016128774584781822 0 +367 1 9.040358 0.99867666 0.0019104437717653883 1 +369 0 -5.1140847 0.011759748 0.017066275150054256 0 +372 0 -3.9699688 0.028265119 0.041365338113873051 0 +374 0 -4.453028 0.019554092 0.028490056372964728 0 +375 0 -5.935418 0.0062253983 0.0090094240348259531 0 +380 0 -5.935418 0.0062253983 0.0090094240348259531 0 +382 0 -3.5015903 0.0402503 0.059269892820826169 0 +385 0 -3.4273868 0.042550277 0.062731363705997648 0 +386 1 5.084236 0.97168756 0.041435602932387204 1 +390 0 -5.446712 0.009093202 0.013178726659715097 0 +393 0 -5.8810935 0.00649353 0.0093987318033745375 0 +394 0 -4.909887 0.013765577 0.019997486443591825 0 +397 0 -4.48624 0.01906283 0.02776736173275238 0 +400 1 7.313369 0.99491894 0.0073491029866983746 1 +401 0 -4.9692993 0.01314948 0.019096521726067024 0 +402 0 -2.7191267 0.071738176 0.10739630802462989 0 +403 0 -3.7908158 0.032373704 0.047478118696930631 0 +405 0 -5.669884 0.007649496 0.011078316226716699 0 +407 0 -5.669884 0.007649496 0.011078316226716699 0 +408 0 -3.537508 0.039180245 0.057662281285319125 0 +410 0 -5.669884 0.007649496 0.011078316226716699 0 +411 0 ? ? ? 0 +412 1 7.6394253 0.996057 0.0056998288375171091 1 +417 0 -5.669884 0.007649496 0.011078316226716699 0 +420 0 -2.696971 0.07289934 0.109202109310663 0 +421 1 9.498289 0.9990743 0.0013361498208689703 1 +424 0 -4.9692993 0.01314948 0.019096521726067024 0 +425 1 11.849485 0.99985236 0.00021301623841437668 1 +426 0 -2.2324486 0.10155222 0.15449343881561414 0 +427 1 4.1596622 0.9433984 0.084060916290148127 1 +431 0 -2.9302087 0.061503273 0.09157638044705417 0 +432 0 -3.5334377 0.03930012 0.057842289520505352 0 +433 0 -4.013695 0.027341869 0.039995277533374687 0 +435 1 7.0216722 0.99362683 0.009223960722536808 1 +437 0 -4.48624 0.01906283 0.02776736173275238 0 +438 0 -3.5384207 0.039153416 0.057621996617189197 0 +443 0 -5.54259 0.008442539 0.012231715171451501 0 +444 0 -2.6761098 0.07400846 0.11092907791467586 0 +445 0 -5.4523587 0.00905354 0.013120983569838577 0 +446 0 -5.935418 0.0062253983 0.0090094240348259531 0 +447 0 -4.0031805 0.027561158 0.040320574900445029 0 +448 0 -5.8810935 0.00649353 0.0093987318033745375 0 +458 0 -3.7524433 0.033326026 0.048898694820578771 0 +459 0 -3.5017056 0.040246822 0.059264662567010976 0 +460 0 -3.5568357 0.038615786 0.05681497968492219 0 +461 0 -3.3979883 0.043495804 0.06415679758406917 0 +462 0 -3.0737762 0.055339187 0.082131682799238909 0 +463 0 -4.282031 0.022286423 0.032516207331833705 0 +468 0 -4.48624 0.01906283 0.02776736173275238 0 +469 0 -5.419147 0.009289299 0.013464259841054335 0 +470 0 -4.649441 0.016819376 0.024471611493406355 0 +471 0 -3.0737762 0.055339187 0.082131682799238909 0 +472 0 -3.4657965 0.041344486 0.060915608927740013 0 +473 0 -4.48624 0.01906283 0.02776736173275238 0 +475 0 -4.9692993 0.01314948 0.019096521726067024 0 +476 0 -4.2355022 0.023092354 0.033705914010540669 0 +477 0 -4.48624 0.01906283 0.02776736173275238 0 +478 0 -3.744658 0.033522516 0.049191973216976458 0 +479 1 6.673234 0.9916496 0.012097621138980599 1 +481 0 -2.5658808 0.08013057 0.12049900060468864 0 +485 0 -4.6490927 0.016823873 0.024478209452408795 0 +486 0 -4.649441 0.016819376 0.024471611493406355 0 +488 1 0.95910263 0.57765627 0.79171681395149796 1 +490 0 -5.935418 0.0062253983 0.0090094240348259531 0 +491 1 5.556223 0.98024607 0.028784146529073261 1 +494 0 -0.015696526 0.3897512 0.7125305321347537 0 +496 0 -5.8810935 0.00649353 0.0093987318033745375 0 +498 0 -4.220706 0.023354555 0.034093183590717085 0 +499 0 -4.220706 0.023354555 0.034093183590717085 0 +500 0 -2.9890532 0.05890275 0.087584283244111522 0 +503 0 -3.955172 0.028584344 0.041839357341208416 0 +505 0 -4.4522543 0.019565687 0.028507118098448581 0 +506 1 8.766859 0.99836195 0.0023651512585342507 1 +508 0 -4.0031805 0.027561158 0.040320574900445029 0 +509 0 -5.4523587 0.00905354 0.013120983569838577 0 +511 0 -3.7376466 0.03370043 0.0494575738020104 0 +512 0 -4.0031805 0.027561158 0.040320574900445029 0 +515 1 7.3749876 0.9951566 0.0070045457965879953 1 +516 0 -5.8810935 0.00649353 0.0093987318033745375 0 +518 0 -4.49539 0.018929625 0.027571465962874155 0 +524 0 -4.7037654 0.016131792 0.023463019884524334 0 +525 0 -4.746127 0.015614876 0.022705238295456929 0 +526 0 -4.48624 0.01906283 0.02776736173275238 0 +530 1 5.325205 0.9764312 0.034409713804049308 1 +536 0 -3.4721127 0.04114936 0.060621989293231218 0 +537 0 -3.2213755 0.04961142 0.07341059341561007 0 +542 0 -3.4861355 0.040719297 0.059975058703050588 0 +543 0 -4.220706 0.023354555 0.034093183590717085 0 +545 0 -3.7376466 0.03370043 0.0494575738020104 0 +550 0 -4.7037654 0.016131792 0.023463019884524334 0 +551 0 -5.4043503 0.009396285 0.013620063190565418 0 +552 0 -3.2360563 0.04907345 0.072594184325982886 0 +553 0 -1.4225531 0.17546193 0.27834199385026431 0 +554 0 -4.9692993 0.01314948 0.019096521726067024 0 +555 0 -1.7431297 0.14211453 0.22114304631062356 0 +556 0 -2.9508896 0.060577318 0.090153667664306186 0 +562 0 -5.4043503 0.009396285 0.013620063190565418 0 +564 0 -3.7509632 0.033363298 0.04895432111111532 0 +567 0 -3.4350505 0.042307038 0.062364895082251487 0 +568 1 3.5748348 0.91345596 0.1305929153487409 1 +570 1 6.466877 0.99020326 0.014203393857934222 1 +571 1 9.048693 0.99868524 0.0018980446658527542 1 +572 0 -4.7037654 0.016131792 0.023463019884524334 0 +573 0 -5.669884 0.007649496 0.011078316226716699 0 +574 1 5.533702 0.9799025 0.029289877586105746 1 +575 0 -3.2213755 0.04961142 0.07341059341561007 0 +576 0 -3.7376466 0.03370043 0.0494575738020104 0 +579 0 -5.4043503 0.009396285 0.013620063190565418 0 +580 0 -3.4869094 0.040695686 0.059939549915604975 0 +583 0 -4.9692993 0.01314948 0.019096521726067024 0 +585 0 -5.935418 0.0062253983 0.0090094240348259531 0 +587 0 -3.5334377 0.03930012 0.057842289520505352 0 +588 1 4.6442137 0.9605316 0.058095029247836778 1 +589 0 -4.0031805 0.027561158 0.040320574900445029 0 +591 1 4.243067 0.9467788 0.078900732087929318 1 +592 1 4.8517904 0.96624035 0.049545999536437443 1 +595 0 -3.7376466 0.03370043 0.0494575738020104 0 +596 0 -3.9699688 0.028265119 0.041365338113873051 0 +597 0 -2.9706378 0.059705306 0.088815117706894264 0 +598 0 -4.7037654 0.016131792 0.023463019884524334 0 +599 0 -2.9381208 0.061147466 0.09102952445231563 0 +601 0 -5.6155596 0.007978472 0.011556665699955653 0 +603 1 3.1762495 0.8854622 0.1754973347343691 1 +605 1 8.159748 0.99737054 0.0037985035173251067 1 +608 1 8.079367 0.9972006 0.0040443320699722306 1 +610 1 6.972575 0.9933793 0.0095834180703139316 1 +611 1 5.494137 0.9792847 0.030199744813973507 1 +615 0 -3.7192311 0.03417207 0.05016190956415939 0 +616 0 -4.7037654 0.016131792 0.023463019884524334 0 +620 0 -4.7037654 0.016131792 0.023463019884524334 0 +623 0 -5.935418 0.0062253983 0.0090094240348259531 0 +625 0 -3.343666 0.045295946 0.066874509831910151 0 +626 1 3.8647957 0.92976606 0.10506033360236412 1 +628 0 -5.4523587 0.00905354 0.013120983569838577 0 +630 0 -2.7601237 0.069634475 0.10413045633568027 0 +631 0 -3.7376466 0.03370043 0.0494575738020104 0 +632 0 -5.935418 0.0062253983 0.0090094240348259531 0 +635 0 -4.2170873 0.023419123 0.034188567100013993 0 +636 1 8.162586 0.9973763 0.0037901403924413767 1 +637 0 -2.370799 0.092107005 0.13940582521487643 0 +640 0 -4.039895 0.026802776 0.039195890391283787 0 +643 0 -5.935418 0.0062253983 0.0090094240348259531 0 +646 0 -5.426158 0.009239029 0.013391057403945885 0 +647 0 -5.4890733 0.008799811 0.012751632164192488 0 +648 1 8.579456 0.9981042 0.0027376359023654488 1 +650 0 -3.6219683 0.036770534 0.054048568515248828 0 +651 0 -4.965017 0.013192964 0.019160092388520032 0 +655 0 -4.7037654 0.016131792 0.023463019884524334 0 +658 1 7.546403 0.99576104 0.006128529583862446 1 +659 0 -5.935418 0.0062253983 0.0090094240348259531 0 +662 0 -5.1684093 0.011276631 0.016361162924390743 0 +663 0 -5.1684093 0.011276631 0.016361162924390743 0 +664 0 -4.2574205 0.02270922 0.033140214869637266 0 +666 0 -3.0702114 0.05548495 0.082354310867892036 0 +667 0 -4.4382315 0.019776938 0.028818005735429076 0 +669 1 6.9817877 0.9934265 0.0095148606603544482 1 +671 0 -3.9565368 0.028554756 0.041795415305621239 0 +672 0 -4.921291 0.013645152 0.019821334809691313 0 +673 0 -3.289723 0.047153607 0.069684436821807083 0 +674 0 -5.669884 0.007649496 0.011078316226716699 0 +675 0 -3.7340279 0.03379261 0.049595209486744243 0 +676 0 -5.419147 0.009289299 0.013464259841054335 0 +677 0 -4.0031805 0.027561158 0.040320574900445029 0 +684 0 -5.935418 0.0062253983 0.0090094240348259531 0 +686 0 -5.935418 0.0062253983 0.0090094240348259531 0 +687 0 -4.295347 0.022060875 0.032183431935772396 0 +690 0 -5.4890733 0.008799811 0.012751632164192488 0 +695 0 -5.4523587 0.00905354 0.013120983569838577 0 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-out.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-out.txt new file mode 100644 index 0000000000..08ac6f6825 --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-out.txt @@ -0,0 +1,39 @@ +maml.exe TrainTest test=%Data% tr=AveragedPerceptron cali=PAV dout=%Output% data=%Data% out=%Output% seed=1 +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Warning: Skipped 160 instances with missing features during training (over 10 iterations; 16 inst/iter) +Training calibrator. +PAV calibrator: piecewise function approximation has 9 components. +Warning: The predictor produced non-finite prediction values on 16 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3499 (239.0/(239.0+444.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 234 | 5 | 0.9791 + negative || 12 | 432 | 0.9730 + ||====================== +Precision || 0.9512 | 0.9886 | +OVERALL 0/1 ACCURACY: 0.975110 +LOG LOSS/instance: 0.084507 +Test-set entropy (prior Log-Loss/instance): 0.934003 +LOG-LOSS REDUCTION (RIG): 0.909522 +AUC: 0.996146 + +OVERALL RESULTS +--------------------------------------- +AUC: 0.996146 (0.0000) +Accuracy: 0.975110 (0.0000) +Positive precision: 0.951220 (0.0000) +Positive recall: 0.979079 (0.0000) +Negative precision: 0.988558 (0.0000) +Negative recall: 0.972973 (0.0000) +Log-loss: 0.084507 (0.0000) +Log-loss reduction: 0.909522 (0.0000) +F1 Score: 0.964948 (0.0000) +AUPRC: 0.992065 (0.0000) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-rp.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-rp.txt new file mode 100644 index 0000000000..e852d92116 --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration-rp.txt @@ -0,0 +1,4 @@ +AveragedPerceptron +AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.996146 0.97511 0.95122 0.979079 0.988558 0.972973 0.084507 0.909522 0.964948 0.992065 AveragedPerceptron %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=AveragedPerceptron cali=PAV dout=%Output% data=%Data% out=%Output% seed=1 + diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration.txt new file mode 100644 index 0000000000..2aaae1e094 --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.PAVcalibration.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +0 0 -3.5726995 1E-15 1.4415419267167138E-15 0 +1 0 3.6101456 0.8333333 2.5849623287385155 1 +2 0 -4.070944 1E-15 1.4415419267167138E-15 0 +3 0 2.470541 0.8095238 2.3923175087700885 1 +4 0 -3.4358397 1E-15 1.4415419267167138E-15 0 +5 1 12.382595 1 -0 1 +6 0 -1.4209604 0.071428575 0.10691520887754996 0 +7 0 -4.7010875 1E-15 1.4415419267167138E-15 0 +8 0 -4.6745405 1E-15 1.4415419267167138E-15 0 +9 0 -4.406417 1E-15 1.4415419267167138E-15 0 +10 0 -5.5593443 1E-15 1.4415419267167138E-15 0 +11 0 -5.4818344 1E-15 1.4415419267167138E-15 0 +12 1 -0.14206791 0.6363636 0.65207672114864346 0 +13 0 -4.569188 1E-15 1.4415419267167138E-15 0 +14 1 9.321613 1 -0 1 +15 1 1.3856993 0.8095238 0.30485456129516797 1 +16 0 -4.533843 1E-15 1.4415419267167138E-15 0 +17 0 -4.046695 1E-15 1.4415419267167138E-15 0 +18 1 7.8903713 1 -0 1 +19 0 -3.0987039 1E-15 1.4415419267167138E-15 0 +20 1 7.528511 1 -0 1 +21 1 7.875206 1 -0 1 +22 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +23 1 ? ? ? 0 +24 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +25 1 1.7418289 0.8095238 0.30485456129516797 1 +26 0 -4.971074 1E-15 1.4415419267167138E-15 0 +27 0 -4.0598474 1E-15 1.4415419267167138E-15 0 +28 0 -5.4818344 1E-15 1.4415419267167138E-15 0 +29 0 -5.8557806 1E-15 1.4415419267167138E-15 0 +30 0 -5.0985003 1E-15 1.4415419267167138E-15 0 +31 0 -4.9946866 1E-15 1.4415419267167138E-15 0 +32 1 7.46414 1 -0 1 +33 0 -4.689259 1E-15 1.4415419267167138E-15 0 +34 0 -4.71424 1E-15 1.4415419267167138E-15 0 +35 0 -5.4818344 1E-15 1.4415419267167138E-15 0 +36 1 9.09911 1 -0 1 +37 0 -1.113348 0.071428575 0.10691520887754996 0 +38 1 6.140954 0.98 0.029146317580716615 1 +39 1 2.5109034 0.8095238 0.30485456129516797 1 +40 0 ? ? ? 0 +41 1 3.3300762 0.8333333 0.26303444023032446 1 +42 1 8.577511 1 -0 1 +43 1 0.49126053 0.6363636 0.65207672114864346 1 +44 1 8.255749 1 -0 1 +45 0 -5.63222 1E-15 1.4415419267167138E-15 0 +46 1 4.5673857 0.9285714 0.10691524360481655 1 +47 0 -5.95583 1E-15 1.4415419267167138E-15 0 +48 0 -3.4358397 1E-15 1.4415419267167138E-15 0 +49 1 5.3666544 0.98 0.029146317580716615 1 +50 1 2.5949688 0.8095238 0.30485456129516797 1 +51 1 0.12595844 0.6363636 0.65207672114864346 1 +52 1 5.2992115 0.98 0.029146317580716615 1 +53 1 8.407227 1 -0 1 +54 1 7.649309 1 -0 1 +55 1 4.478711 0.9285714 0.10691524360481655 1 +56 1 5.5541325 0.98 0.029146317580716615 1 +57 1 1.6657066 0.8095238 0.30485456129516797 1 +58 1 2.5265903 0.8095238 0.30485456129516797 1 +59 1 1.7368546 0.8095238 0.30485456129516797 1 +60 1 2.3288136 0.8095238 0.30485456129516797 1 +61 0 -5.5060835 1E-15 1.4415419267167138E-15 0 +62 1 6.380089 0.98 0.029146317580716615 1 +63 1 0.33490086 0.6363636 0.65207672114864346 1 +64 0 -5.95583 1E-15 1.4415419267167138E-15 0 +65 1 3.8072643 0.9285714 0.10691524360481655 1 +66 0 -4.046695 1E-15 1.4415419267167138E-15 0 +67 1 4.218014 0.9285714 0.10691524360481655 1 +68 1 10.826725 1 -0 1 +69 0 -5.2716546 1E-15 1.4415419267167138E-15 0 +70 0 -3.4726496 1E-15 1.4415419267167138E-15 0 +71 1 7.895046 1 -0 1 +72 0 -2.1755848 0.071428575 0.10691520887754996 0 +73 1 8.9055195 1 -0 1 +74 1 2.5993576 0.8095238 0.30485456129516797 1 +75 0 -4.0411606 1E-15 1.4415419267167138E-15 0 +76 0 -5.0750337 1E-15 1.4415419267167138E-15 0 +77 0 -3.499567 1E-15 1.4415419267167138E-15 0 +78 0 -3.6211972 1E-15 1.4415419267167138E-15 0 +79 0 -5.3911724 1E-15 1.4415419267167138E-15 0 +80 0 -2.7157316 1E-15 1.4415419267167138E-15 0 +81 0 -4.2284155 1E-15 1.4415419267167138E-15 0 +82 0 -3.4452734 1E-15 1.4415419267167138E-15 0 +83 0 -2.1223526 0.071428575 0.10691520887754996 0 +84 1 9.694054 1 -0 1 +85 1 6.2895594 0.98 0.029146317580716615 1 +86 1 2.6168842 0.8095238 0.30485456129516797 1 +87 1 6.91914 1 -0 1 +88 0 -4.046695 1E-15 1.4415419267167138E-15 0 +89 0 -5.085745 1E-15 1.4415419267167138E-15 0 +90 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +91 0 -5.189559 1E-15 1.4415419267167138E-15 0 +92 0 -4.046695 1E-15 1.4415419267167138E-15 0 +93 0 -5.95583 1E-15 1.4415419267167138E-15 0 +94 0 -4.9946866 1E-15 1.4415419267167138E-15 0 +95 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +96 0 -5.663555 1E-15 1.4415419267167138E-15 0 +97 0 -3.5726995 1E-15 1.4415419267167138E-15 0 +98 1 8.590233 1 -0 1 +99 1 10.917194 1 -0 1 +100 1 4.8476696 0.9285714 0.10691524360481655 1 +101 1 -0.84280396 0.5 1 0 +102 0 -3.753096 1E-15 1.4415419267167138E-15 0 +103 1 1.7746449 0.8095238 0.30485456129516797 1 +104 1 12.140858 1 -0 1 +105 1 2.5560703 0.8095238 0.30485456129516797 1 +106 1 9.259369 1 -0 1 +107 1 6.720646 0.98 0.029146317580716615 1 +108 0 -5.5617743 1E-15 1.4415419267167138E-15 0 +109 1 6.871727 1 -0 1 +110 0 -2.766693 1E-15 1.4415419267167138E-15 0 +111 1 3.848032 0.9285714 0.10691524360481655 1 +112 1 9.425768 1 -0 1 +113 1 9.506622 1 -0 1 +114 0 -3.0727453 1E-15 1.4415419267167138E-15 0 +115 0 -4.6439905 1E-15 1.4415419267167138E-15 0 +116 0 -0.66188717 0.5 1 0 +117 1 9.617275 1 -0 1 +118 0 -5.3621607 1E-15 1.4415419267167138E-15 0 +119 0 -3.9435177 1E-15 1.4415419267167138E-15 0 +120 0 -4.8696556 1E-15 1.4415419267167138E-15 0 +121 0 -3.469522 1E-15 1.4415419267167138E-15 0 +122 1 9.680523 1 -0 1 +123 1 3.8165932 0.9285714 0.10691524360481655 1 +124 1 7.6522446 1 -0 1 +125 0 -5.95583 1E-15 1.4415419267167138E-15 0 +126 1 8.564953 1 -0 1 +127 0 -4.520691 1E-15 1.4415419267167138E-15 0 +128 1 4.84898 0.9285714 0.10691524360481655 1 +129 0 -5.717684 1E-15 1.4415419267167138E-15 0 +130 0 -3.4726496 1E-15 1.4415419267167138E-15 0 +131 0 -4.9946866 1E-15 1.4415419267167138E-15 0 +132 1 8.602232 1 -0 1 +133 0 -4.810811 1E-15 1.4415419267167138E-15 0 +134 0 -4.9171767 1E-15 1.4415419267167138E-15 0 +135 0 -2.7288966 1E-15 1.4415419267167138E-15 0 +136 0 -4.533843 1E-15 1.4415419267167138E-15 0 +137 0 -5.494987 1E-15 1.4415419267167138E-15 0 +138 0 -4.2402444 1E-15 1.4415419267167138E-15 0 +139 0 ? ? ? 0 +140 0 -5.494987 1E-15 1.4415419267167138E-15 0 +141 0 -5.9689827 1E-15 1.4415419267167138E-15 0 +142 1 4.4324026 0.9285714 0.10691524360481655 1 +143 0 -4.6439905 1E-15 1.4415419267167138E-15 0 +144 0 -5.4818344 1E-15 1.4415419267167138E-15 0 +145 0 ? ? ? 0 +146 1 1.3394346 0.8095238 0.30485456129516797 1 +147 0 -5.4154215 1E-15 1.4415419267167138E-15 0 +148 0 -1.012373 0.071428575 0.10691520887754996 0 +149 1 11.461615 1 -0 1 +150 0 -5.5593443 1E-15 1.4415419267167138E-15 0 +151 1 5.006484 0.9285714 0.10691524360481655 1 +152 1 9.715748 1 -0 1 +153 0 -4.121497 1E-15 1.4415419267167138E-15 0 +154 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +155 1 3.7769232 0.9285714 0.10691524360481655 1 +156 0 -5.5348053 1E-15 1.4415419267167138E-15 0 +157 0 -4.9946866 1E-15 1.4415419267167138E-15 0 +158 0 ? ? ? 0 +159 1 12.346203 1 -0 1 +160 1 9.039494 1 -0 1 +161 0 -3.8496675 1E-15 1.4415419267167138E-15 0 +162 0 -4.520691 1E-15 1.4415419267167138E-15 0 +163 0 -3.3870554 1E-15 1.4415419267167138E-15 0 +164 0 ? ? ? 0 +165 0 -3.3999205 1E-15 1.4415419267167138E-15 0 +166 1 7.976185 1 -0 1 +167 1 8.355644 1 -0 1 +168 0 -4.520691 1E-15 1.4415419267167138E-15 0 +169 0 -6.2282124 1E-15 1.4415419267167138E-15 0 +170 0 -5.494987 1E-15 1.4415419267167138E-15 0 +171 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +172 0 -5.95583 1E-15 1.4415419267167138E-15 0 +173 1 15.1560135 1 -0 1 +174 1 6.1769257 0.98 0.029146317580716615 1 +175 1 7.842922 1 -0 1 +176 0 -4.9946866 1E-15 1.4415419267167138E-15 0 +177 1 4.766121 0.9285714 0.10691524360481655 1 +178 0 -4.046695 1E-15 1.4415419267167138E-15 0 +179 1 2.290575 0.8095238 0.30485456129516797 1 +180 0 -5.5593443 1E-15 1.4415419267167138E-15 0 +181 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +182 0 -3.0987039 1E-15 1.4415419267167138E-15 0 +183 1 9.1599655 1 -0 1 +184 1 6.2014637 0.98 0.029146317580716615 1 +185 0 -5.085348 1E-15 1.4415419267167138E-15 0 +186 1 5.7654095 0.98 0.029146317580716615 1 +187 1 13.977451 1 -0 1 +188 1 9.065283 1 -0 1 +189 0 -4.7540584 1E-15 1.4415419267167138E-15 0 +190 1 11.957218 1 -0 1 +191 1 10.956873 1 -0 1 +192 0 -4.0598474 1E-15 1.4415419267167138E-15 0 +193 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +194 0 -4.520691 1E-15 1.4415419267167138E-15 0 +195 0 -4.046695 1E-15 1.4415419267167138E-15 0 +196 0 6.8652763 0.98 5.6438575656365879 1 +197 0 -2.6564164 1E-15 1.4415419267167138E-15 0 +198 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +199 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +200 1 10.36586 1 -0 1 +201 1 9.869495 1 -0 1 +202 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +203 0 -3.5726995 1E-15 1.4415419267167138E-15 0 +204 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +205 1 12.086601 1 -0 1 +206 1 5.94417 0.98 0.029146317580716615 1 +207 0 -5.5593443 1E-15 1.4415419267167138E-15 0 +208 0 -5.5593443 1E-15 1.4415419267167138E-15 0 +209 0 -3.6633615 1E-15 1.4415419267167138E-15 0 +210 1 14.534115 1 -0 1 +211 1 9.64962 1 -0 1 +212 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +213 1 14.529058 1 -0 1 +214 1 13.868912 1 -0 1 +215 1 7.643734 1 -0 1 +216 0 -5.95583 1E-15 1.4415419267167138E-15 0 +217 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +218 1 7.88678 1 -0 1 +219 0 -2.511506 1E-15 1.4415419267167138E-15 0 +220 0 -5.1632547 1E-15 1.4415419267167138E-15 0 +221 1 10.395218 1 -0 1 +222 1 -2.2146616 0.071428575 3.8073548575641118 0 +223 1 5.7424126 0.98 0.029146317580716615 1 +224 1 9.995327 1 -0 1 +225 0 -5.95583 1E-15 1.4415419267167138E-15 0 +226 1 10.225868 1 -0 1 +227 1 7.459606 1 -0 1 +228 0 -5.5593443 1E-15 1.4415419267167138E-15 0 +229 1 12.666515 1 -0 1 +230 1 6.1583214 0.98 0.029146317580716615 1 +231 1 8.623032 1 -0 1 +232 0 1.2822819 0.6363636 1.4594315756416352 1 +233 1 6.382519 0.98 0.029146317580716615 1 +234 0 -2.8964386 1E-15 1.4415419267167138E-15 0 +235 0 ? ? ? 0 +236 1 11.420414 1 -0 1 +237 1 6.535795 0.98 0.029146317580716615 1 +238 1 12.422876 1 -0 1 +239 1 5.902529 0.98 0.029146317580716615 1 +240 0 -2.0179915 0.071428575 0.10691520887754996 0 +241 0 -4.000497 1E-15 1.4415419267167138E-15 0 +242 0 -4.9946866 1E-15 1.4415419267167138E-15 0 +243 0 -2.6953983 1E-15 1.4415419267167138E-15 0 +244 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +245 0 -2.817525 1E-15 1.4415419267167138E-15 0 +246 1 11.424002 1 -0 1 +247 1 3.104394 0.8333333 0.26303444023032446 1 +248 0 -3.0615559 1E-15 1.4415419267167138E-15 0 +249 0 ? ? ? 0 +250 0 -6.021953 1E-15 1.4415419267167138E-15 0 +251 1 8.872498 1 -0 1 +252 0 4.5387735 0.9285714 3.8073544061097437 1 +253 1 8.577511 1 -0 1 +254 1 6.380089 0.98 0.029146317580716615 1 +255 1 4.052038 0.9285714 0.10691524360481655 1 +256 0 -5.494987 1E-15 1.4415419267167138E-15 0 +257 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +258 0 -4.520691 1E-15 1.4415419267167138E-15 0 +259 0 2.9647484 0.8095238 2.3923175087700885 1 +260 1 9.870926 1 -0 1 +261 1 12.206299 1 -0 1 +262 1 9.653839 1 -0 1 +263 1 8.981979 1 -0 1 +264 1 5.664709 0.98 0.029146317580716615 1 +265 0 -2.494875 1E-15 1.4415419267167138E-15 0 +266 1 7.3661633 1 -0 1 +267 1 3.3009605 0.8333333 0.26303444023032446 1 +268 1 9.372967 1 -0 1 +269 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +270 1 6.031378 0.98 0.029146317580716615 1 +271 0 -3.5726995 1E-15 1.4415419267167138E-15 0 +272 1 3.3009605 0.8333333 0.26303444023032446 1 +273 1 0.21747303 0.6363636 0.65207672114864346 1 +274 0 -4.323663 1E-15 1.4415419267167138E-15 0 +275 0 ? ? ? 0 +276 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +277 0 -5.95583 1E-15 1.4415419267167138E-15 0 +278 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +279 1 7.127905 1 -0 1 +280 0 -4.520691 1E-15 1.4415419267167138E-15 0 +281 0 -4.689259 1E-15 1.4415419267167138E-15 0 +282 1 4.4381237 0.9285714 0.10691524360481655 1 +283 1 6.0636253 0.98 0.029146317580716615 1 +284 1 7.431343 1 -0 1 +285 1 14.218479 1 -0 1 +286 1 15.281265 1 -0 1 +287 0 -4.9171767 1E-15 1.4415419267167138E-15 0 +288 1 2.2163525 0.8095238 0.30485456129516797 1 +289 1 8.312019 1 -0 1 +290 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +291 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +292 1 ? ? ? 0 +293 1 5.542122 0.98 0.029146317580716615 1 +294 0 ? ? ? 0 +295 1 7.7866364 1 -0 1 +296 0 1.823431 0.8095238 2.3923175087700885 1 +297 0 ? ? ? 0 +298 0 -2.725597 1E-15 1.4415419267167138E-15 0 +299 1 7.8274345 1 -0 1 +300 1 7.348076 1 -0 1 +301 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +302 1 15.735764 1 -0 1 +303 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +304 1 5.9607983 0.98 0.029146317580716615 1 +305 1 8.459471 1 -0 1 +306 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +307 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +308 1 7.422592 1 -0 1 +309 0 -1.7474074 0.071428575 0.10691520887754996 0 +310 0 -5.3911724 1E-15 1.4415419267167138E-15 0 +311 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +312 1 3.6294708 0.9285714 0.10691524360481655 1 +313 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +314 0 -6.0464916 1E-15 1.4415419267167138E-15 0 +315 0 ? ? ? 0 +316 1 3.6177406 0.9285714 0.10691524360481655 1 +317 1 9.215706 1 -0 1 +318 0 -5.1966968 1E-15 1.4415419267167138E-15 0 +319 0 2.6369457 0.8095238 2.3923175087700885 1 +320 1 7.3824844 1 -0 1 +321 0 ? ? ? 0 +322 0 -4.520691 1E-15 1.4415419267167138E-15 0 +323 1 5.5612926 0.98 0.029146317580716615 1 +324 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +325 0 -4.192758 1E-15 1.4415419267167138E-15 0 +326 1 4.4103794 0.9285714 0.10691524360481655 1 +327 0 -5.95583 1E-15 1.4415419267167138E-15 0 +328 1 3.373888 0.8333333 0.26303444023032446 1 +329 1 7.8321342 1 -0 1 +330 1 5.856251 0.98 0.029146317580716615 1 +331 0 -3.2490892 1E-15 1.4415419267167138E-15 0 +332 0 -3.1363668 1E-15 1.4415419267167138E-15 0 +333 1 4.914962 0.9285714 0.10691524360481655 1 +334 1 5.9119453 0.98 0.029146317580716615 1 +335 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +336 1 5.543519 0.98 0.029146317580716615 1 +337 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +338 0 -6.0464916 1E-15 1.4415419267167138E-15 0 +339 1 5.684025 0.98 0.029146317580716615 1 +340 1 6.620781 0.98 0.029146317580716615 1 +341 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +342 0 -5.9689827 1E-15 1.4415419267167138E-15 0 +343 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +344 1 10.162451 1 -0 1 +345 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +346 0 -3.335825 1E-15 1.4415419267167138E-15 0 +347 0 -6.1395845 1E-15 1.4415419267167138E-15 0 +348 1 0.15727425 0.6363636 0.65207672114864346 1 +349 1 4.0622606 0.9285714 0.10691524360481655 1 +350 0 -3.93614 1E-15 1.4415419267167138E-15 0 +351 0 -4.9946866 1E-15 1.4415419267167138E-15 0 +352 0 0.47192955 0.6363636 1.4594315756416352 1 +353 1 8.696344 1 -0 1 +354 0 -5.95583 1E-15 1.4415419267167138E-15 0 +355 0 -4.2461534 1E-15 1.4415419267167138E-15 0 +356 1 -0.69921684 0.5 1 0 +357 1 12.852016 1 -0 1 +358 1 5.582206 0.98 0.029146317580716615 1 +359 1 5.3672857 0.98 0.029146317580716615 1 +360 1 15.333874 1 -0 1 +361 1 6.317689 0.98 0.029146317580716615 1 +362 0 -3.5059962 1E-15 1.4415419267167138E-15 0 +363 0 -2.0658464 0.071428575 0.10691520887754996 0 +364 0 -4.9946866 1E-15 1.4415419267167138E-15 0 +365 0 -5.4818344 1E-15 1.4415419267167138E-15 0 +366 1 13.694571 1 -0 1 +367 1 11.299244 1 -0 1 +368 0 -5.8557806 1E-15 1.4415419267167138E-15 0 +369 0 -5.4592943 1E-15 1.4415419267167138E-15 0 +370 0 -3.89473 1E-15 1.4415419267167138E-15 0 +371 0 -5.8557806 1E-15 1.4415419267167138E-15 0 +372 0 -4.2402444 1E-15 1.4415419267167138E-15 0 +373 0 -3.7544198 1E-15 1.4415419267167138E-15 0 +374 0 -4.71424 1E-15 1.4415419267167138E-15 0 +375 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +376 0 -5.95583 1E-15 1.4415419267167138E-15 0 +377 0 -6.0464916 1E-15 1.4415419267167138E-15 0 +378 0 -3.803866 1E-15 1.4415419267167138E-15 0 +379 0 -2.2557268 1E-15 1.4415419267167138E-15 0 +380 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +381 1 10.07641 1 -0 1 +382 0 -3.5972104 1E-15 1.4415419267167138E-15 0 +383 0 -5.9689827 1E-15 1.4415419267167138E-15 0 +384 0 -5.9689827 1E-15 1.4415419267167138E-15 0 +385 0 -3.7061968 1E-15 1.4415419267167138E-15 0 +386 1 6.0875864 0.98 0.029146317580716615 1 +387 0 -2.3345594 1E-15 1.4415419267167138E-15 0 +388 0 -5.2848067 1E-15 1.4415419267167138E-15 0 +389 0 -3.322411 1E-15 1.4415419267167138E-15 0 +390 0 -5.6504025 1E-15 1.4415419267167138E-15 0 +391 1 10.030338 1 -0 1 +392 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +393 0 -6.53364 1E-15 1.4415419267167138E-15 0 +394 0 -5.241206 1E-15 1.4415419267167138E-15 0 +395 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +396 0 -4.520691 1E-15 1.4415419267167138E-15 0 +397 0 -5.0209913 1E-15 1.4415419267167138E-15 0 +398 0 -4.683385 1E-15 1.4415419267167138E-15 0 +399 0 -5.7283545 1E-15 1.4415419267167138E-15 0 +400 1 10.056744 1 -0 1 +401 0 -5.494987 1E-15 1.4415419267167138E-15 0 +402 0 -3.2177973 1E-15 1.4415419267167138E-15 0 +403 0 -4.145746 1E-15 1.4415419267167138E-15 0 +404 0 -5.507669 1E-15 1.4415419267167138E-15 0 +405 0 -5.95583 1E-15 1.4415419267167138E-15 0 +406 0 -4.1128182 1E-15 1.4415419267167138E-15 0 +407 0 -5.95583 1E-15 1.4415419267167138E-15 0 +408 0 -3.6910605 1E-15 1.4415419267167138E-15 0 +409 0 -4.71424 1E-15 1.4415419267167138E-15 0 +410 0 -5.95583 1E-15 1.4415419267167138E-15 0 +411 0 ? ? ? 0 +412 1 9.230705 1 -0 1 +413 0 -3.2791004 1E-15 1.4415419267167138E-15 0 +414 1 6.7173824 0.98 0.029146317580716615 1 +415 0 -0.6668339 0.5 1 0 +416 1 8.809383 1 -0 1 +417 0 -5.95583 1E-15 1.4415419267167138E-15 0 +418 0 -1.8758612 0.071428575 0.10691520887754996 0 +419 0 -5.442145 1E-15 1.4415419267167138E-15 0 +420 0 -2.5893164 1E-15 1.4415419267167138E-15 0 +421 1 11.824856 1 -0 1 +422 0 -2.8105888 1E-15 1.4415419267167138E-15 0 +423 0 -3.4726496 1E-15 1.4415419267167138E-15 0 +424 0 -5.494987 1E-15 1.4415419267167138E-15 0 +425 1 14.817663 1 -0 1 +426 0 -2.8241482 1E-15 1.4415419267167138E-15 0 +427 1 4.3530827 0.9285714 0.10691524360481655 1 +428 0 -5.95583 1E-15 1.4415419267167138E-15 0 +429 0 -5.4818344 1E-15 1.4415419267167138E-15 0 +430 0 -5.50395 1E-15 1.4415419267167138E-15 0 +431 0 -2.5834928 1E-15 1.4415419267167138E-15 0 +432 0 -3.8628197 1E-15 1.4415419267167138E-15 0 +433 0 -4.463106 1E-15 1.4415419267167138E-15 0 +434 0 5.0084 0.9285714 3.8073544061097437 1 +435 1 7.444332 1 -0 1 +436 1 3.841199 0.9285714 0.10691524360481655 1 +437 0 -5.0209913 1E-15 1.4415419267167138E-15 0 +438 0 -3.822938 1E-15 1.4415419267167138E-15 0 +439 0 -4.5469956 1E-15 1.4415419267167138E-15 0 +440 1 10.154686 1 -0 1 +441 0 -1.8604469 0.071428575 0.10691520887754996 0 +442 0 -4.9326286 1E-15 1.4415419267167138E-15 0 +443 0 -5.9932313 1E-15 1.4415419267167138E-15 0 +444 0 -2.442047 1E-15 1.4415419267167138E-15 0 +445 0 -5.9689827 1E-15 1.4415419267167138E-15 0 +446 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +447 0 -4.5469956 1E-15 1.4415419267167138E-15 0 +448 0 -6.53364 1E-15 1.4415419267167138E-15 0 +449 1 10.298004 1 -0 1 +450 0 -4.0136495 1E-15 1.4415419267167138E-15 0 +451 0 -4.5469956 1E-15 1.4415419267167138E-15 0 +452 0 -4.8841314 1E-15 1.4415419267167138E-15 0 +453 1 8.777971 1 -0 1 +454 0 -5.6234684 1E-15 1.4415419267167138E-15 0 +455 1 0.81635284 0.6363636 0.65207672114864346 1 +456 1 10.487385 1 -0 1 +457 1 9.06263 1 -0 1 +458 0 -4.2533965 1E-15 1.4415419267167138E-15 0 +459 0 -3.9597979 1E-15 1.4415419267167138E-15 0 +460 0 -3.93614 1E-15 1.4415419267167138E-15 0 +461 0 -3.6959996 1E-15 1.4415419267167138E-15 0 +462 0 -3.4621444 1E-15 1.4415419267167138E-15 0 +463 0 -4.823963 1E-15 1.4415419267167138E-15 0 +464 0 -5.0209913 1E-15 1.4415419267167138E-15 0 +465 1 9.7853985 1 -0 1 +466 1 9.541931 1 -0 1 +467 1 7.7145195 1 -0 1 +468 0 -5.0209913 1E-15 1.4415419267167138E-15 0 +469 0 -5.6622314 1E-15 1.4415419267167138E-15 0 +470 0 -5.0985003 1E-15 1.4415419267167138E-15 0 +471 0 -3.4621444 1E-15 1.4415419267167138E-15 0 +472 0 -4.163662 1E-15 1.4415419267167138E-15 0 +473 0 -5.0209913 1E-15 1.4415419267167138E-15 0 +474 0 -4.5469956 1E-15 1.4415419267167138E-15 0 +475 0 -5.494987 1E-15 1.4415419267167138E-15 0 +476 0 -4.727392 1E-15 1.4415419267167138E-15 0 +477 0 -5.0209913 1E-15 1.4415419267167138E-15 0 +478 0 -4.4195695 1E-15 1.4415419267167138E-15 0 +479 1 8.32148 1 -0 1 +480 0 -4.6376576 1E-15 1.4415419267167138E-15 0 +481 0 -3.082261 1E-15 1.4415419267167138E-15 0 +482 1 15.481422 1 -0 1 +483 1 10.906593 1 -0 1 +484 0 -4.2533965 1E-15 1.4415419267167138E-15 0 +485 0 -5.0232906 1E-15 1.4415419267167138E-15 0 +486 0 -5.0985003 1E-15 1.4415419267167138E-15 0 +487 1 13.475904 1 -0 1 +488 1 1.3273249 0.8095238 0.30485456129516797 1 +489 1 -0.6232014 0.5 1 0 +490 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +491 1 6.6713333 0.98 0.029146317580716615 1 +492 0 -4.624505 1E-15 1.4415419267167138E-15 0 +493 1 9.94891 1 -0 1 +494 0 0.9629116 0.6363636 1.4594315756416352 1 +495 0 -5.0985003 1E-15 1.4415419267167138E-15 0 +496 0 -6.53364 1E-15 1.4415419267167138E-15 0 +497 0 -4.893565 1E-15 1.4415419267167138E-15 0 +498 0 -4.533843 1E-15 1.4415419267167138E-15 0 +499 0 -4.533843 1E-15 1.4415419267167138E-15 0 +500 0 -3.0987039 1E-15 1.4415419267167138E-15 0 +501 0 -4.533843 1E-15 1.4415419267167138E-15 0 +502 0 -4.2284155 1E-15 1.4415419267167138E-15 0 +503 0 -4.046695 1E-15 1.4415419267167138E-15 0 +504 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +505 0 -5.2401347 1E-15 1.4415419267167138E-15 0 +506 1 10.447666 1 -0 1 +507 0 -5.0937376 1E-15 1.4415419267167138E-15 0 +508 0 -4.5469956 1E-15 1.4415419267167138E-15 0 +509 0 -5.9689827 1E-15 1.4415419267167138E-15 0 +510 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +511 0 -4.0598474 1E-15 1.4415419267167138E-15 0 +512 0 -4.5469956 1E-15 1.4415419267167138E-15 0 +513 0 -5.0985003 1E-15 1.4415419267167138E-15 0 +514 1 10.719854 1 -0 1 +515 1 8.6480255 1 -0 1 +516 0 -6.53364 1E-15 1.4415419267167138E-15 0 +517 0 -6.0464916 1E-15 1.4415419267167138E-15 0 +518 0 -4.8959603 1E-15 1.4415419267167138E-15 0 +519 1 6.535844 0.98 0.029146317580716615 1 +520 0 -6.3523164 1E-15 1.4415419267167138E-15 0 +521 0 -4.9303293 1E-15 1.4415419267167138E-15 0 +522 1 5.502533 0.98 0.029146317580716615 1 +523 1 7.699238 1 -0 1 +524 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +525 0 -5.189559 1E-15 1.4415419267167138E-15 0 +526 0 -5.0209913 1E-15 1.4415419267167138E-15 0 +527 0 -4.046695 1E-15 1.4415419267167138E-15 0 +528 0 -3.1803741 1E-15 1.4415419267167138E-15 0 +529 0 -4.624505 1E-15 1.4415419267167138E-15 0 +530 1 6.515936 0.98 0.029146317580716615 1 +531 0 -4.1128182 1E-15 1.4415419267167138E-15 0 +532 0 -5.5593443 1E-15 1.4415419267167138E-15 0 +533 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +534 0 -5.4818344 1E-15 1.4415419267167138E-15 0 +535 0 -4.5760565 1E-15 1.4415419267167138E-15 0 +536 0 -3.5726995 1E-15 1.4415419267167138E-15 0 +537 0 -3.2791004 1E-15 1.4415419267167138E-15 0 +538 0 -4.533843 1E-15 1.4415419267167138E-15 0 +539 0 -3.5858517 1E-15 1.4415419267167138E-15 0 +540 0 -3.6101003 1E-15 1.4415419267167138E-15 0 +541 0 -5.494987 1E-15 1.4415419267167138E-15 0 +542 0 -4.292143 1E-15 1.4415419267167138E-15 0 +543 0 -4.533843 1E-15 1.4415419267167138E-15 0 +544 0 -4.589209 1E-15 1.4415419267167138E-15 0 +545 0 -4.0598474 1E-15 1.4415419267167138E-15 0 +546 1 11.390259 1 -0 1 +547 0 -6.059644 1E-15 1.4415419267167138E-15 0 +548 0 -5.5856485 1E-15 1.4415419267167138E-15 0 +549 1 6.3187475 0.98 0.029146317580716615 1 +550 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +551 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +552 0 -3.1100621 1E-15 1.4415419267167138E-15 0 +553 0 -1.7353673 0.071428575 0.10691520887754996 0 +554 0 -5.494987 1E-15 1.4415419267167138E-15 0 +555 0 -1.925499 0.071428575 0.10691520887754996 0 +556 0 -3.424057 1E-15 1.4415419267167138E-15 0 +557 0 -3.93614 1E-15 1.4415419267167138E-15 0 +558 0 -5.4818344 1E-15 1.4415419267167138E-15 0 +559 0 -4.0598474 1E-15 1.4415419267167138E-15 0 +560 0 -3.5726995 1E-15 1.4415419267167138E-15 0 +561 0 -3.5726995 1E-15 1.4415419267167138E-15 0 +562 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +563 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +564 0 -3.8496675 1E-15 1.4415419267167138E-15 0 +565 1 12.086092 1 -0 1 +566 0 -4.227092 1E-15 1.4415419267167138E-15 0 +567 0 -3.6343489 1E-15 1.4415419267167138E-15 0 +568 1 4.1473064 0.9285714 0.10691524360481655 1 +569 1 10.713882 1 -0 1 +570 1 8.017664 1 -0 1 +571 1 11.034657 1 -0 1 +572 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +573 0 -5.95583 1E-15 1.4415419267167138E-15 0 +574 1 5.950966 0.98 0.029146317580716615 1 +575 0 -3.2791004 1E-15 1.4415419267167138E-15 0 +576 0 -4.0598474 1E-15 1.4415419267167138E-15 0 +577 0 -5.95583 1E-15 1.4415419267167138E-15 0 +578 0 -5.95583 1E-15 1.4415419267167138E-15 0 +579 0 -5.4686823 1E-15 1.4415419267167138E-15 0 +580 0 -3.7662487 1E-15 1.4415419267167138E-15 0 +581 1 8.417797 1 -0 1 +582 1 7.907978 1 -0 1 +583 0 -5.494987 1E-15 1.4415419267167138E-15 0 +584 0 -2.929195 1E-15 1.4415419267167138E-15 0 +585 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +586 1 13.98102 1 -0 1 +587 0 -3.8628197 1E-15 1.4415419267167138E-15 0 +588 1 5.463169 0.98 0.029146317580716615 1 +589 0 -4.5469956 1E-15 1.4415419267167138E-15 0 +590 1 3.9684029 0.9285714 0.10691524360481655 1 +591 1 5.966526 0.98 0.029146317580716615 1 +592 1 5.7801704 0.98 0.029146317580716615 1 +593 0 -4.2533965 1E-15 1.4415419267167138E-15 0 +594 1 5.101776 0.98 0.029146317580716615 1 +595 0 -4.0598474 1E-15 1.4415419267167138E-15 0 +596 0 -4.2402444 1E-15 1.4415419267167138E-15 0 +597 0 -2.9855018 1E-15 1.4415419267167138E-15 0 +598 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +599 0 -3.6294346 1E-15 1.4415419267167138E-15 0 +600 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +601 0 -6.0464916 1E-15 1.4415419267167138E-15 0 +602 0 -4.533843 1E-15 1.4415419267167138E-15 0 +603 1 4.8058815 0.9285714 0.10691524360481655 1 +604 1 6.1928043 0.98 0.029146317580716615 1 +605 1 9.95545 1 -0 1 +606 0 -4.715564 1E-15 1.4415419267167138E-15 0 +607 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +608 1 11.148428 1 -0 1 +609 0 -4.5469956 1E-15 1.4415419267167138E-15 0 +610 1 8.926189 1 -0 1 +611 1 6.9109592 1 -0 1 +612 1 16.893515 1 -0 1 +613 0 -5.226982 1E-15 1.4415419267167138E-15 0 +614 0 -5.5724964 1E-15 1.4415419267167138E-15 0 +615 0 -3.9466453 1E-15 1.4415419267167138E-15 0 +616 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +617 0 ? ? ? 0 +618 0 -4.533843 1E-15 1.4415419267167138E-15 0 +619 0 -4.0598474 1E-15 1.4415419267167138E-15 0 +620 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +621 0 0.3560543 0.6363636 1.4594315756416352 1 +622 0 -2.2074018 0.071428575 0.10691520887754996 0 +623 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +624 0 -3.8450818 1E-15 1.4415419267167138E-15 0 +625 0 -3.4678864 1E-15 1.4415419267167138E-15 0 +626 1 5.7601643 0.98 0.029146317580716615 1 +627 0 -4.1699953 1E-15 1.4415419267167138E-15 0 +628 0 -5.9689827 1E-15 1.4415419267167138E-15 0 +629 0 -5.0209913 1E-15 1.4415419267167138E-15 0 +630 0 -3.3583755 1E-15 1.4415419267167138E-15 0 +631 0 -4.0598474 1E-15 1.4415419267167138E-15 0 +632 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +633 1 4.3299503 0.9285714 0.10691524360481655 1 +634 0 -5.494987 1E-15 1.4415419267167138E-15 0 +635 0 -4.61419 1E-15 1.4415419267167138E-15 0 +636 1 10.12752 1 -0 1 +637 0 -2.4650178 1E-15 1.4415419267167138E-15 0 +638 0 -5.0209913 1E-15 1.4415419267167138E-15 0 +639 0 -3.93614 1E-15 1.4415419267167138E-15 0 +640 0 -4.4101357 1E-15 1.4415419267167138E-15 0 +641 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +642 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +643 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +644 0 -5.9689827 1E-15 1.4415419267167138E-15 0 +645 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +646 0 -6.021953 1E-15 1.4415419267167138E-15 0 +647 0 -5.832123 1E-15 1.4415419267167138E-15 0 +648 1 12.236198 1 -0 1 +649 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +650 0 -3.7496567 1E-15 1.4415419267167138E-15 0 +651 0 -5.2175484 1E-15 1.4415419267167138E-15 0 +652 0 -3.8628197 1E-15 1.4415419267167138E-15 0 +653 0 -4.533843 1E-15 1.4415419267167138E-15 0 +654 0 -4.520691 1E-15 1.4415419267167138E-15 0 +655 0 -5.0078387 1E-15 1.4415419267167138E-15 0 +656 0 -4.0598474 1E-15 1.4415419267167138E-15 0 +657 0 -0.4869156 0.5 1 0 +658 1 9.086258 1 -0 1 +659 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +660 0 -5.95583 1E-15 1.4415419267167138E-15 0 +661 0 -4.046695 1E-15 1.4415419267167138E-15 0 +662 0 -5.3686323 1E-15 1.4415419267167138E-15 0 +663 0 -5.3686323 1E-15 1.4415419267167138E-15 0 +664 0 -4.396983 1E-15 1.4415419267167138E-15 0 +665 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +666 0 -3.496932 1E-15 1.4415419267167138E-15 0 +667 0 -4.520691 1E-15 1.4415419267167138E-15 0 +668 1 2.8043652 0.8095238 0.30485456129516797 1 +669 1 8.147335 1 -0 1 +670 1 6.4856205 0.98 0.029146317580716615 1 +671 0 -4.087837 1E-15 1.4415419267167138E-15 0 +672 0 -4.9946866 1E-15 1.4415419267167138E-15 0 +673 0 -3.9078822 1E-15 1.4415419267167138E-15 0 +674 0 -5.95583 1E-15 1.4415419267167138E-15 0 +675 0 -4.1401944 1E-15 1.4415419267167138E-15 0 +676 0 -5.6622314 1E-15 1.4415419267167138E-15 0 +677 0 -4.5469956 1E-15 1.4415419267167138E-15 0 +678 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +679 0 -5.9689827 1E-15 1.4415419267167138E-15 0 +680 1 16.78001 1 -0 1 +681 1 9.801077 1 -0 1 +682 0 -3.3756719 1E-15 1.4415419267167138E-15 0 +683 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +684 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +685 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +686 0 -6.4429784 1E-15 1.4415419267167138E-15 0 +687 0 -4.613783 1E-15 1.4415419267167138E-15 0 +688 0 -5.0209913 1E-15 1.4415419267167138E-15 0 +689 0 -4.1777043 1E-15 1.4415419267167138E-15 0 +690 0 -5.832123 1E-15 1.4415419267167138E-15 0 +691 1 4.4967794 0.9285714 0.10691524360481655 1 +692 0 -5.494987 1E-15 1.4415419267167138E-15 0 +693 0 -4.6842284 1E-15 1.4415419267167138E-15 0 +694 0 -4.9545784 1E-15 1.4415419267167138E-15 0 +695 0 -5.9689827 1E-15 1.4415419267167138E-15 0 +696 1 6.7127867 0.98 0.029146317580716615 1 +697 1 5.064644 0.98 0.029146317580716615 1 +698 1 6.1036215 0.98 0.029146317580716615 1 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-out.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-out.txt new file mode 100644 index 0000000000..d12e4da2b0 --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-out.txt @@ -0,0 +1,38 @@ +maml.exe TrainTest test=%Data% tr=AveragedPerceptron numcali=200 dout=%Output% data=%Data% out=%Output% seed=1 +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Warning: Skipped 160 instances with missing features during training (over 10 iterations; 16 inst/iter) +Training calibrator. +Warning: The predictor produced non-finite prediction values on 16 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3499 (239.0/(239.0+444.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 234 | 5 | 0.9791 + negative || 12 | 432 | 0.9730 + ||====================== +Precision || 0.9512 | 0.9886 | +OVERALL 0/1 ACCURACY: 0.975110 +LOG LOSS/instance: 0.120617 +Test-set entropy (prior Log-Loss/instance): 0.934003 +LOG-LOSS REDUCTION (RIG): 0.870860 +AUC: 0.996146 + +OVERALL RESULTS +--------------------------------------- +AUC: 0.996146 (0.0000) +Accuracy: 0.975110 (0.0000) +Positive precision: 0.951220 (0.0000) +Positive recall: 0.979079 (0.0000) +Negative precision: 0.988558 (0.0000) +Negative recall: 0.972973 (0.0000) +Log-loss: 0.120617 (0.0000) +Log-loss reduction: 0.870860 (0.0000) +F1 Score: 0.964948 (0.0000) +AUPRC: 0.992065 (0.0000) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-rp.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-rp.txt new file mode 100644 index 0000000000..0860eb7030 --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom-rp.txt @@ -0,0 +1,4 @@ +AveragedPerceptron +AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.996146 0.97511 0.95122 0.979079 0.988558 0.972973 0.120617 0.87086 0.964948 0.992065 AveragedPerceptron %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=AveragedPerceptron numcali=200 dout=%Output% data=%Data% out=%Output% seed=1 + diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom.txt new file mode 100644 index 0000000000..d44554be7c --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.calibrateRandom.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +0 0 -3.5726995 0.04772423 0.070548667598666043 0 +1 0 3.6101456 0.9209522 3.6611308948912158 1 +2 0 -4.070944 0.033202168 0.04871385636169432 0 +3 0 2.470541 0.83073884 2.5626771570538036 1 +4 0 -3.4358397 0.052670263 0.078061422636405919 0 +5 1 12.382595 0.99988943 0.0001595227277818751 1 +6 0 -1.4209604 0.20405397 0.32925748105470037 0 +7 0 -4.7010875 0.020848809 0.030396450646965355 0 +8 0 -4.6745405 0.0212639 0.03100818041936795 0 +9 0 -4.406417 0.02593568 0.03791105404656582 0 +10 0 -5.5593443 0.010982091 0.015931448970605746 0 +11 0 -5.4818344 0.01163943 0.016890638723774297 0 +12 1 -0.14206791 0.40347582 1.3094458747728994 0 +13 0 -4.569188 0.02299235 0.033558236540709968 0 +14 1 9.321613 0.998874 0.0016253773809814112 1 +15 1 1.3856993 0.6830734 0.54988747944718552 1 +16 0 -4.533843 0.023602417 0.03445937066290284 0 +17 0 -4.046695 0.033797726 0.049602846718906586 0 +18 1 7.8903713 0.9966727 0.0048082975514731332 1 +19 0 -3.0987039 0.06699133 0.10003760623051484 0 +20 1 7.528511 0.9956263 0.0063237968594455725 1 +21 1 7.875206 0.9966343 0.0048638619113426269 1 +22 0 -5.0078387 0.016592301 0.024138445598558569 0 +23 1 ? ? ? 0 +24 0 -5.4686823 0.011754765 0.017059001292080284 0 +25 1 1.7418289 0.7384804 0.43736848520073585 1 +26 0 -4.971074 0.017053554 0.024815278480434001 0 +27 0 -4.0598474 0.033473436 0.049118710585029472 0 +28 0 -5.4818344 0.01163943 0.016890638723774297 0 +29 0 -5.8557806 0.008789948 0.012737277035434173 0 +30 0 -5.0985003 0.015506636 0.022546613761495443 0 +31 0 -4.9946866 0.016755883 0.024378445192612154 0 +32 1 7.46414 0.9954084 0.0066395103905325147 1 +33 0 -4.689259 0.021032771 0.030667528713629391 0 +34 0 -4.71424 0.0206461 0.030097808839634525 0 +35 0 -5.4818344 0.01163943 0.016890638723774297 0 +36 1 9.09911 0.99866724 0.0019240484689165348 1 +37 0 -1.113348 0.24456772 0.40462566668624267 0 +38 1 6.140954 0.9875705 0.018044317542393044 1 +39 1 2.5109034 0.83500063 0.26015080155533699 1 +40 0 ? ? ? 0 +41 1 3.3300762 0.90403634 0.14554732340521623 1 +42 1 8.577511 0.99802166 0.0028569650360322531 1 +43 1 0.49126053 0.5223421 0.93693314652059867 1 +44 1 8.255749 0.9974761 0.0036458194620463992 1 +45 0 -5.63222 0.010397596 0.015079090304826438 0 +46 1 4.5673857 0.9601321 0.05869514874509995 1 +47 0 -5.95583 0.008152754 0.011810146123934495 0 +48 0 -3.4358397 0.052670263 0.078061422636405919 0 +49 1 5.3666544 0.9778563 0.032305655935016012 1 +50 1 2.5949688 0.8435999 0.2453691442570726 1 +51 1 0.12595844 0.4532164 1.1417280158533796 1 +52 1 5.2992115 0.976721 0.033981595778575298 1 +53 1 8.407227 0.99774945 0.0032505195711204126 1 +54 1 7.649309 0.9960077 0.0057712269479890379 1 +55 1 4.478711 0.957476 0.062691739926295564 1 +56 1 5.5541325 0.9807353 0.02806428572372743 1 +57 1 1.6657066 0.7271758 0.45962396180677062 1 +58 1 2.5265903 0.83663356 0.25733221946992052 1 +59 1 1.7368546 0.737751 0.43879410985231937 1 +60 1 2.3288136 0.8150797 0.29498697868205404 1 +61 0 -5.5060835 0.011429706 0.016584539459669284 0 +62 1 6.380089 0.98961115 0.015066341261387473 1 +63 1 0.33490086 0.4927046 1.0212051535714572 1 +64 0 -5.95583 0.008152754 0.011810146123934495 0 +65 1 3.8072643 0.93117535 0.1028752252736623 1 +66 0 -4.046695 0.033797726 0.049602846718906586 0 +67 1 4.218014 0.9486548 0.076044929659653121 1 +68 1 10.826725 0.9996402 0.00051913703181331893 1 +69 0 -5.2716546 0.013623919 0.019790279664758565 0 +70 0 -3.4726496 0.05129428 0.075967446798405988 0 +71 1 7.895046 0.99668443 0.0047913008067115544 1 +72 0 -2.1755848 0.12635368 0.19487874790855292 0 +73 1 8.9055195 0.9984568 0.0022281210940979582 1 +74 1 2.5993576 0.8440387 0.2446189028997863 1 +75 0 -4.0411606 0.0339351 0.049807982437482065 0 +76 0 -5.0750337 0.01578075 0.022948361946263734 0 +77 0 -3.499567 0.050309677 0.074470941275211228 0 +78 0 -3.6211972 0.04607982 0.068059539432443916 0 +79 0 -5.3911724 0.012457771 0.018085653607344938 0 +80 0 -2.7157316 0.08759694 0.13225680660583347 0 +81 0 -4.2284155 0.029574297 0.0433103327967286 0 +82 0 -3.4452734 0.052314337 0.077519483155153973 0 +83 0 -2.1223526 0.13087894 0.20237095101991351 0 +84 1 9.694054 0.99915093 0.0012254667568283894 1 +85 1 6.2895594 0.9888809 0.016131360230936344 1 +86 1 2.6168842 0.84578085 0.24164420028856959 1 +87 1 6.91914 0.9930738 0.010027129067340199 1 +88 0 -4.046695 0.033797726 0.049602846718906586 0 +89 0 -5.085745 0.015655046 0.022764112920425308 0 +90 0 -5.4686823 0.011754765 0.017059001292080284 0 +91 0 -5.189559 0.014486659 0.021052694034437186 0 +92 0 -4.046695 0.033797726 0.049602846718906586 0 +93 0 -5.95583 0.008152754 0.011810146123934495 0 +94 0 -4.9946866 0.016755883 0.024378445192612154 0 +95 0 -5.4686823 0.011754765 0.017059001292080284 0 +96 0 -5.663555 0.010155835 0.014726681160470451 0 +97 0 -3.5726995 0.04772423 0.070548667598666043 0 +98 1 8.590233 0.9980406 0.0028295658493644239 1 +99 1 10.917194 0.99966407 0.00048472853254951715 1 +100 1 4.8476696 0.9675202 0.047636346312542086 1 +101 1 -0.84280396 0.28443488 1.8138296783750301 0 +102 0 -3.753096 0.041876137 0.061715920539130963 0 +103 1 1.7746449 0.74325943 0.42806223277958716 1 +104 1 12.140858 0.9998672 0.00019160139559387039 1 +105 1 2.5560703 0.83966714 0.25211056290937062 1 +106 1 9.259369 0.99881965 0.0017038920105930569 1 +107 1 6.720646 0.99195737 0.011649978820645172 1 +108 0 -5.5617743 0.010962087 0.015902269313839958 0 +109 1 6.871727 0.992822 0.010393022735684912 1 +110 0 -2.766693 0.08455607 0.12745657439713692 0 +111 1 3.848032 0.9331311 0.099848310460100273 1 +112 1 9.425768 0.9989595 0.0015019320975399703 1 +113 1 9.506622 0.9990213 0.0014126689718023965 1 +114 0 -3.0727453 0.06823267 0.10195834856678422 0 +115 0 -4.6439905 0.021751594 0.031727240251371532 0 +116 0 -0.66188717 0.31317267 0.54198064343187602 0 +117 1 9.617275 0.9991 0.0012989676266954972 1 +118 0 -5.3621607 0.012731449 0.018485523010614183 0 +119 0 -3.9435177 0.036449008 0.053567077140436584 0 +120 0 -4.8696556 0.01839226 0.02678146906266772 0 +121 0 -3.469522 0.051409855 0.076143215435917344 0 +122 1 9.680523 0.99914217 0.0012381182790972595 1 +123 1 3.8165932 0.9316275 0.10217485034891999 1 +124 1 7.6522446 0.99601656 0.0057583629403995997 1 +125 0 -5.95583 0.008152754 0.011810146123934495 0 +126 1 8.564953 0.9980027 0.0028843647430662487 1 +127 0 -4.520691 0.023833435 0.034800756571360128 0 +128 1 4.84898 0.9675514 0.04758977495588574 1 +129 0 -5.717684 0.009751257 0.014137129433839279 0 +130 0 -3.4726496 0.05129428 0.075967446798405988 0 +131 0 -4.9946866 0.016755883 0.024378445192612154 0 +132 1 8.602232 0.9980583 0.0028039765128859018 1 +133 0 -4.810811 0.019215727 0.027992249801728063 0 +134 0 -4.9171767 0.017752614 0.025841671374291015 0 +135 0 -2.7288966 0.08680204 0.13100046294100468 0 +136 0 -4.533843 0.023602417 0.03445937066290284 0 +137 0 -5.494987 0.011525209 0.016723921499597793 0 +138 0 -4.2402444 0.029317854 0.042929138502563191 0 +139 0 ? ? ? 0 +140 0 -5.494987 0.011525209 0.016723921499597793 0 +141 0 -5.9689827 0.008072469 0.011693372500291397 0 +142 1 4.4324026 0.9560225 0.064883520941486764 1 +143 0 -4.6439905 0.021751594 0.031727240251371532 0 +144 0 -5.4818344 0.01163943 0.016890638723774297 0 +145 0 ? ? ? 0 +146 1 1.3394346 0.67542744 0.56612731001545713 1 +147 0 -5.4154215 0.012233485 0.017758031610787428 0 +148 0 -1.012373 0.2589947 0.43244423064898052 0 +149 1 11.461615 0.99977773 0.00032069729386163213 1 +150 0 -5.5593443 0.010982091 0.015931448970605746 0 +151 1 5.006484 0.9711001 0.042308092038938809 1 +152 1 9.715748 0.99916476 0.0012054999542047689 1 +153 0 -4.121497 0.031992976 0.046910578381358904 0 +154 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +155 1 3.7769232 0.9296856 0.10518519653096636 1 +156 0 -5.5348053 0.011186127 0.016229110022171814 0 +157 0 -4.9946866 0.016755883 0.024378445192612154 0 +158 0 ? ? ? 0 +159 1 12.346203 0.99988633 0.0001639947780942108 1 +160 1 9.039494 0.99860567 0.002012998795947356 1 +161 0 -3.8496675 0.039033543 0.057442021457383786 0 +162 0 -4.520691 0.023833435 0.034800756571360128 0 +163 0 -3.3870554 0.054547623 0.080923303986291273 0 +164 0 ? ? ? 0 +165 0 -3.3999205 0.054046504 0.080158834184346328 0 +166 1 7.976185 0.9968817 0.004505750926832852 1 +167 1 8.355644 0.99765986 0.0033800618776249807 1 +168 0 -4.520691 0.023833435 0.034800756571360128 0 +169 0 -6.2282124 0.006640949 0.0096128184951518834 0 +170 0 -5.494987 0.011525209 0.016723921499597793 0 +171 0 -5.4686823 0.011754765 0.017059001292080284 0 +172 0 -5.95583 0.008152754 0.011810146123934495 0 +173 1 15.1560135 0.9999865 1.9434170443242565E-05 1 +174 1 6.1769257 0.9879011 0.017561488134486943 1 +175 1 7.842922 0.99655116 0.004984229936716684 1 +176 0 -4.9946866 0.016755883 0.024378445192612154 0 +177 1 4.766121 0.96551895 0.050623517901417933 1 +178 0 -4.046695 0.033797726 0.049602846718906586 0 +179 1 2.290575 0.8106676 0.30281765649168213 1 +180 0 -5.5593443 0.010982091 0.015931448970605746 0 +181 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +182 0 -3.0987039 0.06699133 0.10003760623051484 0 +183 1 9.1599655 0.9987273 0.0018372561468384253 1 +184 1 6.2014637 0.98812157 0.017239546569330137 1 +185 0 -5.085348 0.015659684 0.02277091055345919 0 +186 1 5.7654095 0.9835412 0.023942621168576837 1 +187 1 13.977451 0.99996704 4.7553986690113E-05 1 +188 1 9.065283 0.99863267 0.0019739908621602127 1 +189 0 -4.7540584 0.020044114 0.029211289656756057 0 +190 1 11.957218 0.99984735 0.0002202405944450654 1 +191 1 10.956873 0.999674 0.00047036322693564959 1 +192 0 -4.0598474 0.033473436 0.049118710585029472 0 +193 0 -5.4686823 0.011754765 0.017059001292080284 0 +194 0 -4.520691 0.023833435 0.034800756571360128 0 +195 0 -4.046695 0.033797726 0.049602846718906586 0 +196 0 6.8652763 0.992787 7.1151855643261381 1 +197 0 -2.6564164 0.0912605 0.13806130573063377 0 +198 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +199 0 -5.0078387 0.016592301 0.024138445598558569 0 +200 1 10.36586 0.9994898 0.00073627359086373204 1 +201 1 9.869495 0.9992566 0.0010728826448553252 1 +202 0 -5.4686823 0.011754765 0.017059001292080284 0 +203 0 -3.5726995 0.04772423 0.070548667598666043 0 +204 0 -5.4686823 0.011754765 0.017059001292080284 0 +205 1 12.086601 0.99986166 0.00019959967319162043 1 +206 1 5.94417 0.98559827 0.020928376857247459 1 +207 0 -5.5593443 0.010982091 0.015931448970605746 0 +208 0 -5.5593443 0.010982091 0.015931448970605746 0 +209 0 -3.6633615 0.044694107 0.065965330728940397 0 +210 1 14.534115 0.99997836 3.1215188826316377E-05 1 +211 1 9.64962 0.99912184 0.0012674667953664828 1 +212 0 -5.4686823 0.011754765 0.017059001292080284 0 +213 1 14.529058 0.9999783 3.1301182014901005E-05 1 +214 1 13.868912 0.9999642 5.1681712271066226E-05 1 +215 1 7.643734 0.9959909 0.0057955739072098177 1 +216 0 -5.95583 0.008152754 0.011810146123934495 0 +217 0 -5.4686823 0.011754765 0.017059001292080284 0 +218 1 7.88678 0.99666363 0.0048214119279433977 1 +219 0 -2.511506 0.10079573 0.15327920999035266 0 +220 0 -5.1632547 0.014774314 0.021473853979747867 0 +221 1 10.395218 0.999501 0.00072009905980506843 1 +222 1 -2.2146616 0.12311753 3.0218919147627599 0 +223 1 5.7424126 0.9832564 0.024360423750620218 1 +224 1 9.995327 0.99932426 0.00097521318777220637 1 +225 0 -5.95583 0.008152754 0.011810146123934495 0 +226 1 10.225868 0.9994326 0.00081878372812222485 1 +227 1 7.459606 0.9953926 0.0066624033876966004 1 +228 0 -5.5593443 0.010982091 0.015931448970605746 0 +229 1 12.666515 0.9999109 0.00012856275965089447 1 +230 1 6.1583214 0.9877312 0.017809586200967932 1 +231 1 8.623032 0.99808866 0.0027601224428665879 1 +232 0 1.2822819 0.6658516 1.5814390985815938 1 +233 1 6.382519 0.9896301 0.015038709216571163 1 +234 0 -2.8964386 0.07724249 0.1159765264538533 0 +235 0 ? ? ? 0 +236 1 11.420414 0.99977064 0.00033093257283884215 1 +237 1 6.535795 0.9907579 0.01339555438417007 1 +238 1 12.422876 0.9998928 0.00015470668910224405 1 +239 1 5.902529 0.98514295 0.021595017147241997 1 +240 0 -2.0179915 0.14015016 0.21784335765559609 0 +241 0 -4.000497 0.03496103 0.051340892555048936 0 +242 0 -4.9946866 0.016755883 0.024378445192612154 0 +243 0 -2.6953983 0.08883759 0.13421985919941723 0 +244 0 -5.4686823 0.011754765 0.017059001292080284 0 +245 0 -2.817525 0.0816187 0.122834823030523 0 +246 1 11.424002 0.9997713 0.00032998645156814019 1 +247 1 3.104394 0.88812214 0.17116999419202034 1 +248 0 -3.0615559 0.0687743 0.10279721718614883 0 +249 0 ? ? ? 0 +250 0 -6.021953 0.0077569955 0.011234608902224096 0 +251 1 8.872498 0.9984177 0.0022846196981478615 1 +252 0 4.5387735 0.95929295 4.6185774530404009 1 +253 1 8.577511 0.99802166 0.0028569650360322531 1 +254 1 6.380089 0.98961115 0.015066341261387473 1 +255 1 4.052038 0.94216394 0.085949972449708326 1 +256 0 -5.494987 0.011525209 0.016723921499597793 0 +257 0 -5.0078387 0.016592301 0.024138445598558569 0 +258 0 -4.520691 0.023833435 0.034800756571360128 0 +259 0 2.9647484 0.87715596 3.0251002258142448 1 +260 1 9.870926 0.99925745 0.0010716778711863072 1 +261 1 12.206299 0.99987364 0.00018231312886902128 1 +262 1 9.653839 0.99912465 0.0012634216564773898 1 +263 1 8.981979 0.9985436 0.002102643541396716 1 +264 1 5.664709 0.9822578 0.025826400239286989 1 +265 0 -2.494875 0.101944976 0.15512425230176288 0 +266 1 7.3661633 0.99505585 0.0071505858244227481 1 +267 1 3.3009605 0.902103 0.14863591753511315 1 +268 1 9.372967 0.998917 0.0015633090806578933 1 +269 0 -5.4686823 0.011754765 0.017059001292080284 0 +270 1 6.031378 0.9865078 0.01959767586907752 1 +271 0 -3.5726995 0.04772423 0.070548667598666043 0 +272 1 3.3009605 0.902103 0.14863591753511315 1 +273 1 0.21747303 0.47046885 1.0878288936386407 1 +274 0 -4.323663 0.027569665 0.040333195363167151 0 +275 0 ? ? ? 0 +276 0 -5.0078387 0.016592301 0.024138445598558569 0 +277 0 -5.95583 0.008152754 0.011810146123934495 0 +278 0 -5.4686823 0.011754765 0.017059001292080284 0 +279 1 7.127905 0.9940822 0.0085629246022787941 1 +280 0 -4.520691 0.023833435 0.034800756571360128 0 +281 0 -4.689259 0.021032771 0.030667528713629391 0 +282 1 4.4381237 0.9562046 0.064608759119747658 1 +283 1 6.0636253 0.9868295 0.019127222915579972 1 +284 1 7.431343 0.99529326 0.006806421628886848 1 +285 1 14.218479 0.9999725 3.9642545670177728E-05 1 +286 1 15.281265 0.9999877 1.7714321792245208E-05 1 +287 0 -4.9171767 0.017752614 0.025841671374291015 0 +288 1 2.2163525 0.8018743 0.31855203011846706 1 +289 1 8.312019 0.9975813 0.0034936687628036997 1 +290 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +291 0 -5.4686823 0.011754765 0.017059001292080284 0 +292 1 ? ? ? 0 +293 1 5.542122 0.9805624 0.028318669172894068 1 +294 0 ? ? ? 0 +295 1 7.7866364 0.99640125 0.0052012629006476284 1 +296 0 1.823431 0.75025773 2.0014880731401443 1 +297 0 ? ? ? 0 +298 0 -2.725597 0.08700066 0.13131427853570132 0 +299 1 7.8274345 0.9965105 0.0050430801821242759 1 +300 1 7.348076 0.9949879 0.0072491063826461647 1 +301 0 -5.4686823 0.011754765 0.017059001292080284 0 +302 1 15.735764 0.9999913 1.2554788140693439E-05 1 +303 0 -5.4686823 0.011754765 0.017059001292080284 0 +304 1 5.9607983 0.98577625 0.020667878309752356 1 +305 1 8.459471 0.9978367 0.003124349823870973 1 +306 0 -5.4686823 0.011754765 0.017059001292080284 0 +307 0 -5.4686823 0.011754765 0.017059001292080284 0 +308 1 7.422592 0.9952621 0.0068516084791740307 1 +309 0 -1.7474074 0.1667576 0.26319183874281699 0 +310 0 -5.3911724 0.012457771 0.018085653607344938 0 +311 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +312 1 3.6294708 0.9220128 0.11714130634122567 1 +313 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +314 0 -6.0464916 0.0076150284 0.011028207282755094 0 +315 0 ? ? ? 0 +316 1 3.6177406 0.92137057 0.11814658455212051 1 +317 1 9.215706 0.99877995 0.0017612310516048875 1 +318 0 -5.1966968 0.014409559 0.020939830908957652 0 +319 0 2.6369457 0.84775543 2.7155373416013662 1 +320 1 7.3824844 0.9951164 0.0070627872070711771 1 +321 0 ? ? ? 0 +322 0 -4.520691 0.023833435 0.034800756571360128 0 +323 1 5.5612926 0.98083764 0.027913746222999549 1 +324 0 -5.4686823 0.011754765 0.017059001292080284 0 +325 0 -4.192758 0.030360552 0.044479700816540804 0 +326 1 4.4103794 0.95531476 0.065951946768905437 1 +327 0 -5.95583 0.008152754 0.011810146123934495 0 +328 1 3.373888 0.90688115 0.14101459636069669 1 +329 1 7.8321342 0.9965229 0.0050251314657524084 1 +330 1 5.856251 0.98462033 0.022360562903046506 1 +331 0 -3.2490892 0.060203616 0.089579877460783788 0 +332 0 -3.1363668 0.065227516 0.097312827363238 0 +333 1 4.914962 0.9690866 0.04530251936714582 1 +334 1 5.9119453 0.98524714 0.021442445493118613 1 +335 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +336 1 5.543519 0.9805826 0.028288940560717461 1 +337 0 -5.4686823 0.011754765 0.017059001292080284 0 +338 0 -6.0464916 0.0076150284 0.011028207282755094 0 +339 1 5.684025 0.98251134 0.025454033740074194 1 +340 1 6.620781 0.9913299 0.012562837713558307 1 +341 0 -5.4686823 0.011754765 0.017059001292080284 0 +342 0 -5.9689827 0.008072469 0.011693372500291397 0 +343 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +344 1 10.162451 0.99940467 0.00085913711935366719 1 +345 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +346 0 -3.335825 0.05658689 0.08403844784383796 0 +347 0 -6.1395845 0.007099501 0.010278945699523763 0 +348 1 0.15727425 0.45910957 1.1230895756645052 1 +349 1 4.0622606 0.94258505 0.085305293794040227 1 +350 0 -3.93614 0.03664607 0.053862165334169397 0 +351 0 -4.9946866 0.016755883 0.024378445192612154 0 +352 0 0.47192955 0.51868224 1.0549384413455107 1 +353 1 8.696344 0.9981919 0.0026109079692710987 1 +354 0 -5.95583 0.008152754 0.011810146123934495 0 +355 0 -4.2461534 0.02919056 0.042739958980967357 0 +356 1 -0.69921684 0.30711412 1.7031532309414121 0 +357 1 12.852016 0.9999226 0.00011170705633068974 1 +358 1 5.582206 0.9811335 0.027478611445726227 1 +359 1 5.3672857 0.97786665 0.032290354697391166 1 +360 1 15.333874 0.9999882 1.7026382905914664E-05 1 +361 1 6.317689 0.9891131 0.0157926107591757 1 +362 0 -3.5059962 0.050077166 0.074117773145114488 0 +363 0 -2.0658464 0.13583237 0.21061690231070429 0 +364 0 -4.9946866 0.016755883 0.024378445192612154 0 +365 0 -5.4818344 0.01163943 0.016890638723774297 0 +366 1 13.694571 0.9999591 5.899125529457858E-05 1 +367 1 11.299244 0.9997486 0.00036275701338483524 1 +368 0 -5.8557806 0.008789948 0.012737277035434173 0 +369 0 -5.4592943 0.011837781 0.017180198068166948 0 +370 0 -3.89473 0.03777133 0.055548308059238383 0 +371 0 -5.8557806 0.008789948 0.012737277035434173 0 +372 0 -4.2402444 0.029317854 0.042929138502563191 0 +373 0 -3.7544198 0.041835867 0.061655284675024721 0 +374 0 -4.71424 0.0206461 0.030097808839634525 0 +375 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +376 0 -5.95583 0.008152754 0.011810146123934495 0 +377 0 -6.0464916 0.0076150284 0.011028207282755094 0 +378 0 -3.803866 0.040357873 0.059431603257068664 0 +379 0 -2.2557268 0.119793735 0.18408645524883044 0 +380 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +381 1 10.07641 0.99936455 0.00091704491714961699 1 +382 0 -3.5972104 0.046886288 0.069279748027861715 0 +383 0 -5.9689827 0.008072469 0.011693372500291397 0 +384 0 -5.9689827 0.008072469 0.011693372500291397 0 +385 0 -3.7061968 0.04332707 0.063902319636563967 0 +386 1 6.0875864 0.98706365 0.018784981568614377 1 +387 0 -2.3345594 0.11363011 0.17401921865853265 0 +388 0 -5.2848067 0.013490494 0.01959514256851105 0 +389 0 -3.322411 0.05713257 0.084873154256777064 0 +390 0 -5.6504025 0.010256627 0.014873591896799637 0 +391 1 10.030338 0.99934196 0.00094965672085389622 1 +392 0 -5.0078387 0.016592301 0.024138445598558569 0 +393 0 -6.53364 0.0052747834 0.0076300449418954053 0 +394 0 -5.241206 0.013937824 0.020249477094060098 0 +395 0 -5.0078387 0.016592301 0.024138445598558569 0 +396 0 -4.520691 0.023833435 0.034800756571360128 0 +397 0 -5.0209913 0.016430285 0.023900780900640347 0 +398 0 -4.683385 0.021124719 0.03080303743590964 0 +399 0 -5.7283545 0.009673404 0.014023709757288819 0 +400 1 10.056744 0.999355 0.00093081234331069047 1 +401 0 -5.494987 0.011525209 0.016723921499597793 0 +402 0 -3.2177973 0.061560765 0.091664762799479257 0 +403 0 -4.145746 0.031428173 0.04606905550815843 0 +404 0 -5.507669 0.011416124 0.016564719123992665 0 +405 0 -5.95583 0.008152754 0.011810146123934495 0 +406 0 -4.1128182 0.032197498 0.047215425637133698 0 +407 0 -5.95583 0.008152754 0.011810146123934495 0 +408 0 -3.6910605 0.043805502 0.064623991066693184 0 +409 0 -4.71424 0.0206461 0.030097808839634525 0 +410 0 -5.95583 0.008152754 0.011810146123934495 0 +411 0 ? ? ? 0 +412 1 9.230705 0.9987937 0.0017413429278913565 1 +413 0 -3.2791004 0.058928363 0.087623545818194815 0 +414 1 6.7173824 0.9919376 0.011678759699519599 1 +415 0 -0.6668339 0.3123661 0.54028742034074773 0 +416 1 8.809383 0.9983402 0.0023965899325754819 1 +417 0 -5.95583 0.008152754 0.011810146123934495 0 +418 0 -1.8758612 0.15365386 0.24068027528074648 0 +419 0 -5.442145 0.011990929 0.017403807536640805 0 +420 0 -2.5893164 0.09557055 0.14492012404103641 0 +421 1 11.824856 0.99983126 0.00024346198385283565 1 +422 0 -2.8105888 0.08201396 0.12345587604110214 0 +423 0 -3.4726496 0.05129428 0.075967446798405988 0 +424 0 -5.494987 0.011525209 0.016723921499597793 0 +425 1 14.817663 0.9999826 2.5109685538071405E-05 1 +426 0 -2.8241482 0.08124288 0.1222445723919429 0 +427 1 4.3530827 0.9534222 0.068812893604113898 1 +428 0 -5.95583 0.008152754 0.011810146123934495 0 +429 0 -5.4818344 0.01163943 0.016890638723774297 0 +430 0 -5.50395 0.011448006 0.016611246987367219 0 +431 0 -2.5834928 0.095953085 0.14553045171385326 0 +432 0 -3.8628197 0.03866102 0.056882864694895223 0 +433 0 -4.463106 0.02487115 0.036335230970476633 0 +434 0 5.0084 0.97114086 5.1148279570464927 1 +435 1 7.444332 0.9953392 0.0067398103254229808 1 +436 1 3.841199 0.9328069 0.10034962060198541 1 +437 0 -5.0209913 0.016430285 0.023900780900640347 0 +438 0 -3.822938 0.039801262 0.05859505551935467 0 +439 0 -4.5469956 0.023373576 0.034121282067080448 0 +440 1 10.154686 0.99940115 0.00086421363869066531 1 +441 0 -1.8604469 0.15518063 0.24328518697086191 0 +442 0 -4.9326286 0.017549371 0.025543185823416033 0 +443 0 -5.9932313 0.007926504 0.011481091006230973 0 +444 0 -2.442047 0.10567282 0.16112537467750074 0 +445 0 -5.9689827 0.008072469 0.011693372500291397 0 +446 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +447 0 -4.5469956 0.023373576 0.034121282067080448 0 +448 0 -6.53364 0.0052747834 0.0076300449418954053 0 +449 1 10.298004 0.99946284 0.00077516203531090419 1 +450 0 -4.0136495 0.034625966 0.05084007240381206 0 +451 0 -4.5469956 0.023373576 0.034121282067080448 0 +452 0 -4.8841314 0.018195052 0.02649165697739319 0 +453 1 8.777971 0.99830025 0.0024543010625143486 1 +454 0 -5.6234684 0.010466127 0.015179001086720274 0 +455 1 0.81635284 0.58322746 0.77786945830087184 1 +456 1 10.487385 0.99953467 0.00067149052326743921 1 +457 1 9.06263 0.9986299 0.0019779518845872311 1 +458 0 -4.2533965 0.029035255 0.04250918192930437 0 +459 0 -3.9597979 0.036017753 0.052921517587864737 0 +460 0 -3.93614 0.03664607 0.053862165334169397 0 +461 0 -3.6959996 0.043648835 0.064387632971416475 0 +462 0 -3.4621444 0.051683467 0.076559406503540178 0 +463 0 -4.823963 0.019028595 0.02771701150932674 0 +464 0 -5.0209913 0.016430285 0.023900780900640347 0 +465 1 9.7853985 0.99920774 0.0011434497150273235 1 +466 1 9.541931 0.99904716 0.0013753126588050381 1 +467 1 7.7145195 0.99619967 0.0054931654697981886 1 +468 0 -5.0209913 0.016430285 0.023900780900640347 0 +469 0 -5.6622314 0.010165934 0.014741400881058842 0 +470 0 -5.0985003 0.015506636 0.022546613761495443 0 +471 0 -3.4621444 0.051683467 0.076559406503540178 0 +472 0 -4.163662 0.031017097 0.045456883902242347 0 +473 0 -5.0209913 0.016430285 0.023900780900640347 0 +474 0 -4.5469956 0.023373576 0.034121282067080448 0 +475 0 -5.494987 0.011525209 0.016723921499597793 0 +476 0 -4.727392 0.02044533 0.029802081885974746 0 +477 0 -5.0209913 0.016430285 0.023900780900640347 0 +478 0 -4.4195695 0.025684811 0.037539538394583953 0 +479 1 8.32148 0.9975986 0.0034686710325429922 1 +480 0 -4.6376576 0.02185405 0.031878348552885363 0 +481 0 -3.082261 0.06777517 0.10125015118918052 0 +482 1 15.481422 0.99998945 1.5220544890042414E-05 1 +483 1 10.906593 0.9996614 0.0004885994477486597 1 +484 0 -4.2533965 0.029035255 0.04250918192930437 0 +485 0 -5.0232906 0.016402122 0.023859471861794275 0 +486 0 -5.0985003 0.015506636 0.022546613761495443 0 +487 1 13.475904 0.9999518 6.9568659593352019E-05 1 +488 1 1.3273249 0.67341036 0.57044218705778627 1 +489 1 -0.6232014 0.31951928 1.646025101672679 0 +490 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +491 1 6.6713333 0.9916534 0.012092158077135179 1 +492 0 -4.624505 0.022068352 0.032194461843299216 0 +493 1 9.94891 0.99930006 0.0010101496965434329 1 +494 0 0.9629116 0.6099765 1.3583669326774681 1 +495 0 -5.0985003 0.015506636 0.022546613761495443 0 +496 0 -6.53364 0.0052747834 0.0076300449418954053 0 +497 0 -4.893565 0.018067652 0.026304464528524713 0 +498 0 -4.533843 0.023602417 0.03445937066290284 0 +499 0 -4.533843 0.023602417 0.03445937066290284 0 +500 0 -3.0987039 0.06699133 0.10003760623051484 0 +501 0 -4.533843 0.023602417 0.03445937066290284 0 +502 0 -4.2284155 0.029574297 0.0433103327967286 0 +503 0 -4.046695 0.033797726 0.049602846718906586 0 +504 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +505 0 -5.2401347 0.013948999 0.020265827098814658 0 +506 1 10.447666 0.9995204 0.00069205216452113084 1 +507 0 -5.0937376 0.015561887 0.022627581462809947 0 +508 0 -4.5469956 0.023373576 0.034121282067080448 0 +509 0 -5.9689827 0.008072469 0.011693372500291397 0 +510 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +511 0 -4.0598474 0.033473436 0.049118710585029472 0 +512 0 -4.5469956 0.023373576 0.034121282067080448 0 +513 0 -5.0985003 0.015506636 0.022546613761495443 0 +514 1 10.719854 0.9996099 0.00056292303371696174 1 +515 1 8.6480255 0.9981245 0.0027083436166447333 1 +516 0 -6.53364 0.0052747834 0.0076300449418954053 0 +517 0 -6.0464916 0.0076150284 0.011028207282755094 0 +518 0 -4.8959603 0.018035447 0.026257148207574778 0 +519 1 6.535844 0.9907582 0.013395120416830476 1 +520 0 -6.3523164 0.006047886 0.0087517464841605584 0 +521 0 -4.9303293 0.017579468 0.025587382354727788 0 +522 1 5.502533 0.97998166 0.029173343678465424 1 +523 1 7.699238 0.9961555 0.0055571295392400502 1 +524 0 -5.0078387 0.016592301 0.024138445598558569 0 +525 0 -5.189559 0.014486659 0.021052694034437186 0 +526 0 -5.0209913 0.016430285 0.023900780900640347 0 +527 0 -4.046695 0.033797726 0.049602846718906586 0 +528 0 -3.1803741 0.06322136 0.094219910903214188 0 +529 0 -4.624505 0.022068352 0.032194461843299216 0 +530 1 6.515936 0.9906189 0.013597970981112992 1 +531 0 -4.1128182 0.032197498 0.047215425637133698 0 +532 0 -5.5593443 0.010982091 0.015931448970605746 0 +533 0 -5.0078387 0.016592301 0.024138445598558569 0 +534 0 -5.4818344 0.01163943 0.016890638723774297 0 +535 0 -4.5760565 0.022875601 0.033385850215234145 0 +536 0 -3.5726995 0.04772423 0.070548667598666043 0 +537 0 -3.2791004 0.058928363 0.087623545818194815 0 +538 0 -4.533843 0.023602417 0.03445937066290284 0 +539 0 -3.5858517 0.04727285 0.069864992457484662 0 +540 0 -3.6101003 0.04645126 0.068621411180630684 0 +541 0 -5.494987 0.011525209 0.016723921499597793 0 +542 0 -4.292143 0.028217996 0.041295378130558093 0 +543 0 -4.533843 0.023602417 0.03445937066290284 0 +544 0 -4.589209 0.022653643 0.033058172494570243 0 +545 0 -4.0598474 0.033473436 0.049118710585029472 0 +546 1 11.390259 0.99976534 0.00033858757684800106 1 +547 0 -6.059644 0.0075399983 0.010919135175626083 0 +548 0 -5.5856485 0.010767463 0.015618401982975634 0 +549 1 6.3187475 0.98912174 0.015780004831897315 1 +550 0 -5.0078387 0.016592301 0.024138445598558569 0 +551 0 -5.4686823 0.011754765 0.017059001292080284 0 +552 0 -3.1100621 0.06645479 0.099208204310445816 0 +553 0 -1.7353673 0.16803055 0.26539753298199431 0 +554 0 -5.494987 0.011525209 0.016723921499597793 0 +555 0 -1.925499 0.14882074 0.23246510163403059 0 +556 0 -3.424057 0.053118028 0.078743488261134892 0 +557 0 -3.93614 0.03664607 0.053862165334169397 0 +558 0 -5.4818344 0.01163943 0.016890638723774297 0 +559 0 -4.0598474 0.033473436 0.049118710585029472 0 +560 0 -3.5726995 0.04772423 0.070548667598666043 0 +561 0 -3.5726995 0.04772423 0.070548667598666043 0 +562 0 -5.4686823 0.011754765 0.017059001292080284 0 +563 0 -5.0078387 0.016592301 0.024138445598558569 0 +564 0 -3.8496675 0.039033543 0.057442021457383786 0 +565 1 12.086092 0.9998616 0.00019968567641750592 1 +566 0 -4.227092 0.02960313 0.043353196700953674 0 +567 0 -3.6343489 0.045643263 0.067399450117881038 0 +568 1 4.1473064 0.94597834 0.080120938914498979 1 +569 1 10.713882 0.9996081 0.00056550378256940248 1 +570 1 8.017664 0.99697804 0.0043663610015818432 1 +571 1 11.034657 0.99969274 0.00044335339890107668 1 +572 0 -5.0078387 0.016592301 0.024138445598558569 0 +573 0 -5.95583 0.008152754 0.011810146123934495 0 +574 1 5.950966 0.9856713 0.020821502205009505 1 +575 0 -3.2791004 0.058928363 0.087623545818194815 0 +576 0 -4.0598474 0.033473436 0.049118710585029472 0 +577 0 -5.95583 0.008152754 0.011810146123934495 0 +578 0 -5.95583 0.008152754 0.011810146123934495 0 +579 0 -5.4686823 0.011754765 0.017059001292080284 0 +580 0 -3.7662487 0.041477647 0.061116018080247782 0 +581 1 8.417797 0.9977674 0.0032245780320862543 1 +582 1 7.907978 0.9967167 0.0047446254970359954 1 +583 0 -5.494987 0.011525209 0.016723921499597793 0 +584 0 -2.929195 0.0754899 0.11323901572445229 0 +585 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +586 1 13.98102 0.9999671 4.7467992532759307E-05 1 +587 0 -3.8628197 0.03866102 0.056882864694895223 0 +588 1 5.463169 0.9793874 0.030048455526317758 1 +589 0 -4.5469956 0.023373576 0.034121282067080448 0 +590 1 3.9684029 0.9386082 0.09140498577121392 1 +591 1 5.966526 0.98583704 0.020578904318245012 1 +592 1 5.7801704 0.98372144 0.023678256090487241 1 +593 0 -4.2533965 0.029035255 0.04250918192930437 0 +594 1 5.101776 0.97306114 0.039397632242971693 1 +595 0 -4.0598474 0.033473436 0.049118710585029472 0 +596 0 -4.2402444 0.029317854 0.042929138502563191 0 +597 0 -2.9855018 0.07256242 0.10867790688560501 0 +598 0 -5.0078387 0.016592301 0.024138445598558569 0 +599 0 -3.6294346 0.045805924 0.06764536479405775 0 +600 0 -5.0078387 0.016592301 0.024138445598558569 0 +601 0 -6.0464916 0.0076150284 0.011028207282755094 0 +602 0 -4.533843 0.023602417 0.03445937066290284 0 +603 1 4.8058815 0.9665091 0.049144773305248697 1 +604 1 6.1928043 0.9880442 0.01735250950197232 1 +605 1 9.95545 0.9993035 0.0010051587149210726 1 +606 0 -4.715564 0.020625807 0.030067914584067133 0 +607 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +608 1 11.148428 0.9997181 0.00040671030044872971 1 +609 0 -4.5469956 0.023373576 0.034121282067080448 0 +610 1 8.926189 0.99848074 0.0021934995672910755 1 +611 1 6.9109592 0.993031 0.010089302795775161 1 +612 1 16.893515 0.99999636 5.2454803872987944E-06 1 +613 0 -5.226982 0.014086899 0.020467603236270997 0 +614 0 -5.5724964 0.010874253 0.01577415289360869 0 +615 0 -3.9466453 0.036365777 0.053442464169990524 0 +616 0 -5.0078387 0.016592301 0.024138445598558569 0 +617 0 ? ? ? 0 +618 0 -4.533843 0.023602417 0.03445937066290284 0 +619 0 -4.0598474 0.033473436 0.049118710585029472 0 +620 0 -5.0078387 0.016592301 0.024138445598558569 0 +621 0 0.3560543 0.49671578 0.99055474235503638 1 +622 0 -2.2074018 0.123713315 0.19052515652518898 0 +623 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +624 0 -3.8450818 0.039164234 0.057638240119205257 0 +625 0 -3.4678864 0.05147039 0.076235286523975876 0 +626 1 5.7601643 0.98347664 0.024037311317372478 1 +627 0 -4.1699953 0.030873027 0.045242397512024662 0 +628 0 -5.9689827 0.008072469 0.011693372500291397 0 +629 0 -5.0209913 0.016430285 0.023900780900640347 0 +630 0 -3.3583755 0.05568056 0.082653127317450656 0 +631 0 -4.0598474 0.033473436 0.049118710585029472 0 +632 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +633 1 4.3299503 0.95263666 0.070002027657626642 1 +634 0 -5.494987 0.011525209 0.016723921499597793 0 +635 0 -4.61419 0.022237852 0.032444539675292525 0 +636 1 10.12752 0.9993887 0.00088219670683892097 1 +637 0 -2.4650178 0.10403732 0.15848945806851844 0 +638 0 -5.0209913 0.016430285 0.023900780900640347 0 +639 0 -3.93614 0.03664607 0.053862165334169397 0 +640 0 -4.4101357 0.025864506 0.037805642164735351 0 +641 0 -5.0078387 0.016592301 0.024138445598558569 0 +642 0 -5.0078387 0.016592301 0.024138445598558569 0 +643 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +644 0 -5.9689827 0.008072469 0.011693372500291397 0 +645 0 -5.0078387 0.016592301 0.024138445598558569 0 +646 0 -6.021953 0.0077569955 0.011234608902224096 0 +647 0 -5.832123 0.008947696 0.012966894893009758 0 +648 1 12.236198 0.9998765 0.00017818502951934642 1 +649 0 -5.0078387 0.016592301 0.024138445598558569 0 +650 0 -3.7496567 0.041980945 0.061873742792964384 0 +651 0 -5.2175484 0.014186635 0.020613554183436408 0 +652 0 -3.8628197 0.03866102 0.056882864694895223 0 +653 0 -4.533843 0.023602417 0.03445937066290284 0 +654 0 -4.520691 0.023833435 0.034800756571360128 0 +655 0 -5.0078387 0.016592301 0.024138445598558569 0 +656 0 -4.0598474 0.033473436 0.049118710585029472 0 +657 0 -0.4869156 0.3424043 0.6047272426002388 0 +658 1 9.086258 0.9986542 0.0019429058245893991 1 +659 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +660 0 -5.95583 0.008152754 0.011810146123934495 0 +661 0 -4.046695 0.033797726 0.049602846718906586 0 +662 0 -5.3686323 0.012669889 0.018395568961803433 0 +663 0 -5.3686323 0.012669889 0.018395568961803433 0 +664 0 -4.396983 0.026117083 0.038179756627613311 0 +665 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +666 0 -3.496932 0.050405264 0.074616157195417698 0 +667 0 -4.520691 0.023833435 0.034800756571360128 0 +668 1 2.8043652 0.8634335 0.21184305945125079 1 +669 1 8.147335 0.99726033 0.0039579294672795707 1 +670 1 6.4856205 0.99040276 0.013912762642680866 1 +671 0 -4.087837 0.03279326 0.048103798472176579 0 +672 0 -4.9946866 0.016755883 0.024378445192612154 0 +673 0 -3.9078822 0.03741039 0.055007243111104311 0 +674 0 -5.95583 0.008152754 0.011810146123934495 0 +675 0 -4.1401944 0.03155663 0.046260403573642889 0 +676 0 -5.6622314 0.010165934 0.014741400881058842 0 +677 0 -4.5469956 0.023373576 0.034121282067080448 0 +678 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +679 0 -5.9689827 0.008072469 0.011693372500291397 0 +680 1 16.78001 0.99999607 5.6754386418026423E-06 1 +681 1 9.801077 0.9992171 0.0011299384356403619 1 +682 0 -3.3756719 0.054994684 0.081605649405800548 0 +683 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +684 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +685 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +686 0 -6.4429784 0.0056481967 0.0081717248736384047 0 +687 0 -4.613783 0.022244569 0.032454450245864012 0 +688 0 -5.0209913 0.016430285 0.023900780900640347 0 +689 0 -4.1777043 0.030698536 0.044982664509967324 0 +690 0 -5.832123 0.008947696 0.012966894893009758 0 +691 1 4.4967794 0.9580306 0.061856385651233553 1 +692 0 -5.494987 0.011525209 0.016723921499597793 0 +693 0 -4.6842284 0.02111149 0.030783541008531489 0 +694 0 -4.9545784 0.017264584 0.025125046093120273 0 +695 0 -5.9689827 0.008072469 0.011693372500291397 0 +696 1 6.7127867 0.9919096 0.011719418003861082 1 +697 1 5.064644 0.97231287 0.040507480679364792 1 +698 1 6.1036215 0.9872181 0.018559275661035816 1 diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-out.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-out.txt new file mode 100644 index 0000000000..5d98ed80f7 --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-out.txt @@ -0,0 +1,39 @@ +maml.exe TrainTest test=%Data% tr=AveragedPerceptron cali={} dout=%Output% data=%Data% out=%Output% seed=1 +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Warning: Skipped 160 instances with missing features during training (over 10 iterations; 16 inst/iter) +Not training a calibrator because a valid calibrator trainer was not provided. +Warning: The predictor produced non-finite prediction values on 16 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3499 (239.0/(239.0+444.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 234 | 5 | 0.9791 + negative || 12 | 432 | 0.9730 + ||====================== +Precision || 0.9512 | 0.9886 | +OVERALL 0/1 ACCURACY: 0.975110 +LOG LOSS/instance: NaN +Test-set entropy (prior Log-Loss/instance): 0.934003 +LOG-LOSS REDUCTION (RIG): 0.000000 +AUC: 0.996146 + +OVERALL RESULTS +--------------------------------------- +AUC: 0.996146 (0.0000) +Accuracy: 0.975110 (0.0000) +Positive precision: 0.951220 (0.0000) +Positive recall: 0.979079 (0.0000) +Negative precision: 0.988558 (0.0000) +Negative recall: 0.972973 (0.0000) +Log-loss: NaN (0.0000) +Log-loss reduction: 0.000000 (0.0000) +F1 Score: 0.964948 (0.0000) +AUPRC: 0.992065 (0.0000) + +--------------------------------------- +Warning: Data does not contain a probability column. Will not output the Log-loss column +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-rp.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-rp.txt new file mode 100644 index 0000000000..ac4bbeaf12 --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration-rp.txt @@ -0,0 +1,4 @@ +AveragedPerceptron +AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.996146 0.97511 0.95122 0.979079 0.988558 0.972973 NaN 0 0.964948 0.992065 AveragedPerceptron %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=AveragedPerceptron cali={} dout=%Output% data=%Data% out=%Output% seed=1 + diff --git a/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration.txt b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration.txt new file mode 100644 index 0000000000..933a677e06 --- /dev/null +++ b/test/BaselineOutput/Common/AveragedPerceptron/net9.0/AveragedPerceptron-TrainTest-breast-cancer.nocalibration.txt @@ -0,0 +1,700 @@ +Instance Label Score Assigned +0 0 -3.5726995 0 +1 0 3.6101456 1 +2 0 -4.070944 0 +3 0 2.470541 1 +4 0 -3.4358397 0 +5 1 12.382595 1 +6 0 -1.4209604 0 +7 0 -4.7010875 0 +8 0 -4.6745405 0 +9 0 -4.406417 0 +10 0 -5.5593443 0 +11 0 -5.4818344 0 +12 1 -0.14206791 0 +13 0 -4.569188 0 +14 1 9.321613 1 +15 1 1.3856993 1 +16 0 -4.533843 0 +17 0 -4.046695 0 +18 1 7.8903713 1 +19 0 -3.0987039 0 +20 1 7.528511 1 +21 1 7.875206 1 +22 0 -5.0078387 0 +23 1 ? 0 +24 0 -5.4686823 0 +25 1 1.7418289 1 +26 0 -4.971074 0 +27 0 -4.0598474 0 +28 0 -5.4818344 0 +29 0 -5.8557806 0 +30 0 -5.0985003 0 +31 0 -4.9946866 0 +32 1 7.46414 1 +33 0 -4.689259 0 +34 0 -4.71424 0 +35 0 -5.4818344 0 +36 1 9.09911 1 +37 0 -1.113348 0 +38 1 6.140954 1 +39 1 2.5109034 1 +40 0 ? 0 +41 1 3.3300762 1 +42 1 8.577511 1 +43 1 0.49126053 1 +44 1 8.255749 1 +45 0 -5.63222 0 +46 1 4.5673857 1 +47 0 -5.95583 0 +48 0 -3.4358397 0 +49 1 5.3666544 1 +50 1 2.5949688 1 +51 1 0.12595844 1 +52 1 5.2992115 1 +53 1 8.407227 1 +54 1 7.649309 1 +55 1 4.478711 1 +56 1 5.5541325 1 +57 1 1.6657066 1 +58 1 2.5265903 1 +59 1 1.7368546 1 +60 1 2.3288136 1 +61 0 -5.5060835 0 +62 1 6.380089 1 +63 1 0.33490086 1 +64 0 -5.95583 0 +65 1 3.8072643 1 +66 0 -4.046695 0 +67 1 4.218014 1 +68 1 10.826725 1 +69 0 -5.2716546 0 +70 0 -3.4726496 0 +71 1 7.895046 1 +72 0 -2.1755848 0 +73 1 8.9055195 1 +74 1 2.5993576 1 +75 0 -4.0411606 0 +76 0 -5.0750337 0 +77 0 -3.499567 0 +78 0 -3.6211972 0 +79 0 -5.3911724 0 +80 0 -2.7157316 0 +81 0 -4.2284155 0 +82 0 -3.4452734 0 +83 0 -2.1223526 0 +84 1 9.694054 1 +85 1 6.2895594 1 +86 1 2.6168842 1 +87 1 6.91914 1 +88 0 -4.046695 0 +89 0 -5.085745 0 +90 0 -5.4686823 0 +91 0 -5.189559 0 +92 0 -4.046695 0 +93 0 -5.95583 0 +94 0 -4.9946866 0 +95 0 -5.4686823 0 +96 0 -5.663555 0 +97 0 -3.5726995 0 +98 1 8.590233 1 +99 1 10.917194 1 +100 1 4.8476696 1 +101 1 -0.84280396 0 +102 0 -3.753096 0 +103 1 1.7746449 1 +104 1 12.140858 1 +105 1 2.5560703 1 +106 1 9.259369 1 +107 1 6.720646 1 +108 0 -5.5617743 0 +109 1 6.871727 1 +110 0 -2.766693 0 +111 1 3.848032 1 +112 1 9.425768 1 +113 1 9.506622 1 +114 0 -3.0727453 0 +115 0 -4.6439905 0 +116 0 -0.66188717 0 +117 1 9.617275 1 +118 0 -5.3621607 0 +119 0 -3.9435177 0 +120 0 -4.8696556 0 +121 0 -3.469522 0 +122 1 9.680523 1 +123 1 3.8165932 1 +124 1 7.6522446 1 +125 0 -5.95583 0 +126 1 8.564953 1 +127 0 -4.520691 0 +128 1 4.84898 1 +129 0 -5.717684 0 +130 0 -3.4726496 0 +131 0 -4.9946866 0 +132 1 8.602232 1 +133 0 -4.810811 0 +134 0 -4.9171767 0 +135 0 -2.7288966 0 +136 0 -4.533843 0 +137 0 -5.494987 0 +138 0 -4.2402444 0 +139 0 ? 0 +140 0 -5.494987 0 +141 0 -5.9689827 0 +142 1 4.4324026 1 +143 0 -4.6439905 0 +144 0 -5.4818344 0 +145 0 ? 0 +146 1 1.3394346 1 +147 0 -5.4154215 0 +148 0 -1.012373 0 +149 1 11.461615 1 +150 0 -5.5593443 0 +151 1 5.006484 1 +152 1 9.715748 1 +153 0 -4.121497 0 +154 0 -6.4429784 0 +155 1 3.7769232 1 +156 0 -5.5348053 0 +157 0 -4.9946866 0 +158 0 ? 0 +159 1 12.346203 1 +160 1 9.039494 1 +161 0 -3.8496675 0 +162 0 -4.520691 0 +163 0 -3.3870554 0 +164 0 ? 0 +165 0 -3.3999205 0 +166 1 7.976185 1 +167 1 8.355644 1 +168 0 -4.520691 0 +169 0 -6.2282124 0 +170 0 -5.494987 0 +171 0 -5.4686823 0 +172 0 -5.95583 0 +173 1 15.1560135 1 +174 1 6.1769257 1 +175 1 7.842922 1 +176 0 -4.9946866 0 +177 1 4.766121 1 +178 0 -4.046695 0 +179 1 2.290575 1 +180 0 -5.5593443 0 +181 0 -6.4429784 0 +182 0 -3.0987039 0 +183 1 9.1599655 1 +184 1 6.2014637 1 +185 0 -5.085348 0 +186 1 5.7654095 1 +187 1 13.977451 1 +188 1 9.065283 1 +189 0 -4.7540584 0 +190 1 11.957218 1 +191 1 10.956873 1 +192 0 -4.0598474 0 +193 0 -5.4686823 0 +194 0 -4.520691 0 +195 0 -4.046695 0 +196 0 6.8652763 1 +197 0 -2.6564164 0 +198 0 -6.4429784 0 +199 0 -5.0078387 0 +200 1 10.36586 1 +201 1 9.869495 1 +202 0 -5.4686823 0 +203 0 -3.5726995 0 +204 0 -5.4686823 0 +205 1 12.086601 1 +206 1 5.94417 1 +207 0 -5.5593443 0 +208 0 -5.5593443 0 +209 0 -3.6633615 0 +210 1 14.534115 1 +211 1 9.64962 1 +212 0 -5.4686823 0 +213 1 14.529058 1 +214 1 13.868912 1 +215 1 7.643734 1 +216 0 -5.95583 0 +217 0 -5.4686823 0 +218 1 7.88678 1 +219 0 -2.511506 0 +220 0 -5.1632547 0 +221 1 10.395218 1 +222 1 -2.2146616 0 +223 1 5.7424126 1 +224 1 9.995327 1 +225 0 -5.95583 0 +226 1 10.225868 1 +227 1 7.459606 1 +228 0 -5.5593443 0 +229 1 12.666515 1 +230 1 6.1583214 1 +231 1 8.623032 1 +232 0 1.2822819 1 +233 1 6.382519 1 +234 0 -2.8964386 0 +235 0 ? 0 +236 1 11.420414 1 +237 1 6.535795 1 +238 1 12.422876 1 +239 1 5.902529 1 +240 0 -2.0179915 0 +241 0 -4.000497 0 +242 0 -4.9946866 0 +243 0 -2.6953983 0 +244 0 -5.4686823 0 +245 0 -2.817525 0 +246 1 11.424002 1 +247 1 3.104394 1 +248 0 -3.0615559 0 +249 0 ? 0 +250 0 -6.021953 0 +251 1 8.872498 1 +252 0 4.5387735 1 +253 1 8.577511 1 +254 1 6.380089 1 +255 1 4.052038 1 +256 0 -5.494987 0 +257 0 -5.0078387 0 +258 0 -4.520691 0 +259 0 2.9647484 1 +260 1 9.870926 1 +261 1 12.206299 1 +262 1 9.653839 1 +263 1 8.981979 1 +264 1 5.664709 1 +265 0 -2.494875 0 +266 1 7.3661633 1 +267 1 3.3009605 1 +268 1 9.372967 1 +269 0 -5.4686823 0 +270 1 6.031378 1 +271 0 -3.5726995 0 +272 1 3.3009605 1 +273 1 0.21747303 1 +274 0 -4.323663 0 +275 0 ? 0 +276 0 -5.0078387 0 +277 0 -5.95583 0 +278 0 -5.4686823 0 +279 1 7.127905 1 +280 0 -4.520691 0 +281 0 -4.689259 0 +282 1 4.4381237 1 +283 1 6.0636253 1 +284 1 7.431343 1 +285 1 14.218479 1 +286 1 15.281265 1 +287 0 -4.9171767 0 +288 1 2.2163525 1 +289 1 8.312019 1 +290 0 -6.4429784 0 +291 0 -5.4686823 0 +292 1 ? 0 +293 1 5.542122 1 +294 0 ? 0 +295 1 7.7866364 1 +296 0 1.823431 1 +297 0 ? 0 +298 0 -2.725597 0 +299 1 7.8274345 1 +300 1 7.348076 1 +301 0 -5.4686823 0 +302 1 15.735764 1 +303 0 -5.4686823 0 +304 1 5.9607983 1 +305 1 8.459471 1 +306 0 -5.4686823 0 +307 0 -5.4686823 0 +308 1 7.422592 1 +309 0 -1.7474074 0 +310 0 -5.3911724 0 +311 0 -6.4429784 0 +312 1 3.6294708 1 +313 0 -6.4429784 0 +314 0 -6.0464916 0 +315 0 ? 0 +316 1 3.6177406 1 +317 1 9.215706 1 +318 0 -5.1966968 0 +319 0 2.6369457 1 +320 1 7.3824844 1 +321 0 ? 0 +322 0 -4.520691 0 +323 1 5.5612926 1 +324 0 -5.4686823 0 +325 0 -4.192758 0 +326 1 4.4103794 1 +327 0 -5.95583 0 +328 1 3.373888 1 +329 1 7.8321342 1 +330 1 5.856251 1 +331 0 -3.2490892 0 +332 0 -3.1363668 0 +333 1 4.914962 1 +334 1 5.9119453 1 +335 0 -6.4429784 0 +336 1 5.543519 1 +337 0 -5.4686823 0 +338 0 -6.0464916 0 +339 1 5.684025 1 +340 1 6.620781 1 +341 0 -5.4686823 0 +342 0 -5.9689827 0 +343 0 -6.4429784 0 +344 1 10.162451 1 +345 0 -6.4429784 0 +346 0 -3.335825 0 +347 0 -6.1395845 0 +348 1 0.15727425 1 +349 1 4.0622606 1 +350 0 -3.93614 0 +351 0 -4.9946866 0 +352 0 0.47192955 1 +353 1 8.696344 1 +354 0 -5.95583 0 +355 0 -4.2461534 0 +356 1 -0.69921684 0 +357 1 12.852016 1 +358 1 5.582206 1 +359 1 5.3672857 1 +360 1 15.333874 1 +361 1 6.317689 1 +362 0 -3.5059962 0 +363 0 -2.0658464 0 +364 0 -4.9946866 0 +365 0 -5.4818344 0 +366 1 13.694571 1 +367 1 11.299244 1 +368 0 -5.8557806 0 +369 0 -5.4592943 0 +370 0 -3.89473 0 +371 0 -5.8557806 0 +372 0 -4.2402444 0 +373 0 -3.7544198 0 +374 0 -4.71424 0 +375 0 -6.4429784 0 +376 0 -5.95583 0 +377 0 -6.0464916 0 +378 0 -3.803866 0 +379 0 -2.2557268 0 +380 0 -6.4429784 0 +381 1 10.07641 1 +382 0 -3.5972104 0 +383 0 -5.9689827 0 +384 0 -5.9689827 0 +385 0 -3.7061968 0 +386 1 6.0875864 1 +387 0 -2.3345594 0 +388 0 -5.2848067 0 +389 0 -3.322411 0 +390 0 -5.6504025 0 +391 1 10.030338 1 +392 0 -5.0078387 0 +393 0 -6.53364 0 +394 0 -5.241206 0 +395 0 -5.0078387 0 +396 0 -4.520691 0 +397 0 -5.0209913 0 +398 0 -4.683385 0 +399 0 -5.7283545 0 +400 1 10.056744 1 +401 0 -5.494987 0 +402 0 -3.2177973 0 +403 0 -4.145746 0 +404 0 -5.507669 0 +405 0 -5.95583 0 +406 0 -4.1128182 0 +407 0 -5.95583 0 +408 0 -3.6910605 0 +409 0 -4.71424 0 +410 0 -5.95583 0 +411 0 ? 0 +412 1 9.230705 1 +413 0 -3.2791004 0 +414 1 6.7173824 1 +415 0 -0.6668339 0 +416 1 8.809383 1 +417 0 -5.95583 0 +418 0 -1.8758612 0 +419 0 -5.442145 0 +420 0 -2.5893164 0 +421 1 11.824856 1 +422 0 -2.8105888 0 +423 0 -3.4726496 0 +424 0 -5.494987 0 +425 1 14.817663 1 +426 0 -2.8241482 0 +427 1 4.3530827 1 +428 0 -5.95583 0 +429 0 -5.4818344 0 +430 0 -5.50395 0 +431 0 -2.5834928 0 +432 0 -3.8628197 0 +433 0 -4.463106 0 +434 0 5.0084 1 +435 1 7.444332 1 +436 1 3.841199 1 +437 0 -5.0209913 0 +438 0 -3.822938 0 +439 0 -4.5469956 0 +440 1 10.154686 1 +441 0 -1.8604469 0 +442 0 -4.9326286 0 +443 0 -5.9932313 0 +444 0 -2.442047 0 +445 0 -5.9689827 0 +446 0 -6.4429784 0 +447 0 -4.5469956 0 +448 0 -6.53364 0 +449 1 10.298004 1 +450 0 -4.0136495 0 +451 0 -4.5469956 0 +452 0 -4.8841314 0 +453 1 8.777971 1 +454 0 -5.6234684 0 +455 1 0.81635284 1 +456 1 10.487385 1 +457 1 9.06263 1 +458 0 -4.2533965 0 +459 0 -3.9597979 0 +460 0 -3.93614 0 +461 0 -3.6959996 0 +462 0 -3.4621444 0 +463 0 -4.823963 0 +464 0 -5.0209913 0 +465 1 9.7853985 1 +466 1 9.541931 1 +467 1 7.7145195 1 +468 0 -5.0209913 0 +469 0 -5.6622314 0 +470 0 -5.0985003 0 +471 0 -3.4621444 0 +472 0 -4.163662 0 +473 0 -5.0209913 0 +474 0 -4.5469956 0 +475 0 -5.494987 0 +476 0 -4.727392 0 +477 0 -5.0209913 0 +478 0 -4.4195695 0 +479 1 8.32148 1 +480 0 -4.6376576 0 +481 0 -3.082261 0 +482 1 15.481422 1 +483 1 10.906593 1 +484 0 -4.2533965 0 +485 0 -5.0232906 0 +486 0 -5.0985003 0 +487 1 13.475904 1 +488 1 1.3273249 1 +489 1 -0.6232014 0 +490 0 -6.4429784 0 +491 1 6.6713333 1 +492 0 -4.624505 0 +493 1 9.94891 1 +494 0 0.9629116 1 +495 0 -5.0985003 0 +496 0 -6.53364 0 +497 0 -4.893565 0 +498 0 -4.533843 0 +499 0 -4.533843 0 +500 0 -3.0987039 0 +501 0 -4.533843 0 +502 0 -4.2284155 0 +503 0 -4.046695 0 +504 0 -6.4429784 0 +505 0 -5.2401347 0 +506 1 10.447666 1 +507 0 -5.0937376 0 +508 0 -4.5469956 0 +509 0 -5.9689827 0 +510 0 -6.4429784 0 +511 0 -4.0598474 0 +512 0 -4.5469956 0 +513 0 -5.0985003 0 +514 1 10.719854 1 +515 1 8.6480255 1 +516 0 -6.53364 0 +517 0 -6.0464916 0 +518 0 -4.8959603 0 +519 1 6.535844 1 +520 0 -6.3523164 0 +521 0 -4.9303293 0 +522 1 5.502533 1 +523 1 7.699238 1 +524 0 -5.0078387 0 +525 0 -5.189559 0 +526 0 -5.0209913 0 +527 0 -4.046695 0 +528 0 -3.1803741 0 +529 0 -4.624505 0 +530 1 6.515936 1 +531 0 -4.1128182 0 +532 0 -5.5593443 0 +533 0 -5.0078387 0 +534 0 -5.4818344 0 +535 0 -4.5760565 0 +536 0 -3.5726995 0 +537 0 -3.2791004 0 +538 0 -4.533843 0 +539 0 -3.5858517 0 +540 0 -3.6101003 0 +541 0 -5.494987 0 +542 0 -4.292143 0 +543 0 -4.533843 0 +544 0 -4.589209 0 +545 0 -4.0598474 0 +546 1 11.390259 1 +547 0 -6.059644 0 +548 0 -5.5856485 0 +549 1 6.3187475 1 +550 0 -5.0078387 0 +551 0 -5.4686823 0 +552 0 -3.1100621 0 +553 0 -1.7353673 0 +554 0 -5.494987 0 +555 0 -1.925499 0 +556 0 -3.424057 0 +557 0 -3.93614 0 +558 0 -5.4818344 0 +559 0 -4.0598474 0 +560 0 -3.5726995 0 +561 0 -3.5726995 0 +562 0 -5.4686823 0 +563 0 -5.0078387 0 +564 0 -3.8496675 0 +565 1 12.086092 1 +566 0 -4.227092 0 +567 0 -3.6343489 0 +568 1 4.1473064 1 +569 1 10.713882 1 +570 1 8.017664 1 +571 1 11.034657 1 +572 0 -5.0078387 0 +573 0 -5.95583 0 +574 1 5.950966 1 +575 0 -3.2791004 0 +576 0 -4.0598474 0 +577 0 -5.95583 0 +578 0 -5.95583 0 +579 0 -5.4686823 0 +580 0 -3.7662487 0 +581 1 8.417797 1 +582 1 7.907978 1 +583 0 -5.494987 0 +584 0 -2.929195 0 +585 0 -6.4429784 0 +586 1 13.98102 1 +587 0 -3.8628197 0 +588 1 5.463169 1 +589 0 -4.5469956 0 +590 1 3.9684029 1 +591 1 5.966526 1 +592 1 5.7801704 1 +593 0 -4.2533965 0 +594 1 5.101776 1 +595 0 -4.0598474 0 +596 0 -4.2402444 0 +597 0 -2.9855018 0 +598 0 -5.0078387 0 +599 0 -3.6294346 0 +600 0 -5.0078387 0 +601 0 -6.0464916 0 +602 0 -4.533843 0 +603 1 4.8058815 1 +604 1 6.1928043 1 +605 1 9.95545 1 +606 0 -4.715564 0 +607 0 -6.4429784 0 +608 1 11.148428 1 +609 0 -4.5469956 0 +610 1 8.926189 1 +611 1 6.9109592 1 +612 1 16.893515 1 +613 0 -5.226982 0 +614 0 -5.5724964 0 +615 0 -3.9466453 0 +616 0 -5.0078387 0 +617 0 ? 0 +618 0 -4.533843 0 +619 0 -4.0598474 0 +620 0 -5.0078387 0 +621 0 0.3560543 1 +622 0 -2.2074018 0 +623 0 -6.4429784 0 +624 0 -3.8450818 0 +625 0 -3.4678864 0 +626 1 5.7601643 1 +627 0 -4.1699953 0 +628 0 -5.9689827 0 +629 0 -5.0209913 0 +630 0 -3.3583755 0 +631 0 -4.0598474 0 +632 0 -6.4429784 0 +633 1 4.3299503 1 +634 0 -5.494987 0 +635 0 -4.61419 0 +636 1 10.12752 1 +637 0 -2.4650178 0 +638 0 -5.0209913 0 +639 0 -3.93614 0 +640 0 -4.4101357 0 +641 0 -5.0078387 0 +642 0 -5.0078387 0 +643 0 -6.4429784 0 +644 0 -5.9689827 0 +645 0 -5.0078387 0 +646 0 -6.021953 0 +647 0 -5.832123 0 +648 1 12.236198 1 +649 0 -5.0078387 0 +650 0 -3.7496567 0 +651 0 -5.2175484 0 +652 0 -3.8628197 0 +653 0 -4.533843 0 +654 0 -4.520691 0 +655 0 -5.0078387 0 +656 0 -4.0598474 0 +657 0 -0.4869156 0 +658 1 9.086258 1 +659 0 -6.4429784 0 +660 0 -5.95583 0 +661 0 -4.046695 0 +662 0 -5.3686323 0 +663 0 -5.3686323 0 +664 0 -4.396983 0 +665 0 -6.4429784 0 +666 0 -3.496932 0 +667 0 -4.520691 0 +668 1 2.8043652 1 +669 1 8.147335 1 +670 1 6.4856205 1 +671 0 -4.087837 0 +672 0 -4.9946866 0 +673 0 -3.9078822 0 +674 0 -5.95583 0 +675 0 -4.1401944 0 +676 0 -5.6622314 0 +677 0 -4.5469956 0 +678 0 -6.4429784 0 +679 0 -5.9689827 0 +680 1 16.78001 1 +681 1 9.801077 1 +682 0 -3.3756719 0 +683 0 -6.4429784 0 +684 0 -6.4429784 0 +685 0 -6.4429784 0 +686 0 -6.4429784 0 +687 0 -4.613783 0 +688 0 -5.0209913 0 +689 0 -4.1777043 0 +690 0 -5.832123 0 +691 1 4.4967794 1 +692 0 -5.494987 0 +693 0 -4.6842284 0 +694 0 -4.9545784 0 +695 0 -5.9689827 0 +696 1 6.7127867 1 +697 1 5.064644 1 +698 1 6.1036215 1 diff --git a/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-def-CV-breast-cancer-out.txt b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-def-CV-breast-cancer-out.txt new file mode 100644 index 0000000000..9710e32a48 --- /dev/null +++ b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-def-CV-breast-cancer-out.txt @@ -0,0 +1,56 @@ +maml.exe CV tr=LdSvm{iter=1000} threads=- dout=%Output% data=%Data% seed=1 +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Warning: Skipped 8 rows with missing feature/label values +Training calibrator. +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Warning: Skipped 8 rows with missing feature/label values +Training calibrator. +Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3785 (134.0/(134.0+220.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 134 | 0 | 1.0000 + negative || 9 | 211 | 0.9591 + ||====================== +Precision || 0.9371 | 1.0000 | +OVERALL 0/1 ACCURACY: 0.974576 +LOG LOSS/instance: 0.122048 +Test-set entropy (prior Log-Loss/instance): 0.956998 +LOG-LOSS REDUCTION (RIG): 0.872467 +AUC: 0.994437 +Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3191 (105.0/(105.0+224.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 101 | 4 | 0.9619 + negative || 5 | 219 | 0.9777 + ||====================== +Precision || 0.9528 | 0.9821 | +OVERALL 0/1 ACCURACY: 0.972644 +LOG LOSS/instance: 0.172708 +Test-set entropy (prior Log-Loss/instance): 0.903454 +LOG-LOSS REDUCTION (RIG): 0.808836 +AUC: 0.971386 + +OVERALL RESULTS +--------------------------------------- +AUC: 0.982911 (0.0115) +Accuracy: 0.973610 (0.0010) +Positive precision: 0.944947 (0.0079) +Positive recall: 0.980952 (0.0190) +Negative precision: 0.991031 (0.0090) +Negative recall: 0.968385 (0.0093) +Log-loss: 0.147378 (0.0253) +Log-loss reduction: 0.840652 (0.0318) +F1 Score: 0.962427 (0.0051) +AUPRC: 0.983054 (0.0068) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-def-CV-breast-cancer.txt b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-def-CV-breast-cancer.txt new file mode 100644 index 0000000000..b8e44ccd53 --- /dev/null +++ b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-def-CV-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +5 1 3.1349773 0.99962956 0.00053453510102781496 1 +6 0 -0.9244971 0.036377292 0.053459703645348515 0 +8 0 -1.305758 0.013041346 0.018938447288102071 0 +9 0 -1.4139888 0.009713269 0.014081786128038521 0 +10 0 -1.3678652 0.011014068 0.015978094989696334 0 +11 0 -1.4135613 0.0097245965 0.014098288900044151 0 +18 1 1.8516258 0.9874686 0.0181932210904747 1 +20 1 1.7854059 0.98499984 0.021804611268681436 1 +21 1 2.1207554 0.99398774 0.0087000387423584059 1 +25 1 0.7185533 0.77680665 0.3643725388857616 1 +28 0 -1.4135613 0.0097245965 0.014098288900044151 0 +31 0 -1.3816847 0.010607217 0.01538471954762626 0 +32 1 2.2605245 0.9959004 0.0059266401299438087 1 +35 0 -1.4135613 0.0097245965 0.014098288900044151 0 +37 0 -1.1632192 0.019188995 0.027952927613419058 0 +40 0 ? ? ? 0 +41 1 1.1791247 0.9252055 0.11215420779745383 1 +44 1 2.5638814 0.9982176 0.0025737790518699562 1 +45 0 -1.3784529 0.01070101 0.015521490792765711 0 +46 1 1.4361101 0.961685 0.056363675618763799 1 +48 0 -1.3627725 0.01116786 0.01620245848409705 0 +50 1 1.0287565 0.891025 0.16646217308587058 1 +51 1 0.4828803 0.64525825 0.63205141756137762 1 +52 1 1.4845697 0.9663096 0.049442590131660989 1 +54 1 2.0668023 0.99303156 0.010088523442763611 1 +56 1 1.6220342 0.976678 0.034045074612715227 1 +60 1 1.0073475 0.885168 0.17597677214206878 1 +63 1 0.74138844 0.7875176 0.34461591544641107 1 +64 0 -1.4110423 0.009791615 0.014195929527819996 0 +66 0 -1.3761241 0.010769103 0.015620793844278126 0 +68 1 2.8579056 0.99920595 0.0011460315025583032 1 +69 0 -1.3782976 0.010705538 0.015528094123528379 0 +70 0 -1.3411375 0.011845209 0.017191041799475098 0 +71 1 2.5835383 0.99831134 0.0024382795322685793 1 +72 0 -1.1336713 0.020781599 0.030297426150736759 0 +73 1 2.0137851 0.99194515 0.011667750078647451 1 +74 1 0.95431995 0.86947256 0.2017875917528727 1 +76 0 -1.3245906 0.012390515 0.017987402459155689 0 +77 0 -1.4025875 0.010019919 0.014528597162487571 0 +79 0 -1.426981 0.009375147 0.013589278234321408 0 +82 0 -1.3697402 0.010957974 0.015896270161493271 0 +88 0 -1.3761241 0.010769103 0.015620793844278126 0 +90 0 -1.3817354 0.010605751 0.015382580668868604 0 +91 0 -1.4407243 0.0090301605 0.013086945707172507 0 +92 0 -1.3761241 0.010769103 0.015620793844278126 0 +93 0 -1.4110423 0.009791615 0.014195929527819996 0 +95 0 -1.3817354 0.010605751 0.015382580668868604 0 +96 0 -1.4381903 0.009092808 0.013178153095375312 0 +97 0 -1.3707396 0.010928193 0.015852830032950474 0 +98 1 2.6065261 0.99841475 0.0022888399570760247 1 +99 1 2.6127307 0.9984416 0.0022500829410306207 1 +100 1 1.6583542 0.9788505 0.030839585025205878 1 +102 0 -1.341929 0.01181973 0.017153843041929236 0 +104 1 3.4889104 0.99986017 0.00020174975537682981 1 +105 1 1.0520489 0.8970978 0.15666277931304001 1 +106 1 2.0419948 0.9925426 0.010799035154487786 1 +108 0 -1.4051962 0.009948921 0.01442513640565986 0 +109 1 1.8430048 0.9871715 0.018627393137037741 1 +111 1 1.1348424 0.91631854 0.12607889160781086 1 +112 1 2.021353 0.99210995 0.011428073251990358 1 +113 1 2.7539048 0.9989429 0.0015258627845947082 1 +115 0 -1.2425047 0.015483979 0.022513411834843233 0 +117 1 2.4448397 0.99752796 0.0035708196615732994 1 +120 0 -1.3468829 0.01166147 0.016922810171817185 0 +121 0 -1.3219292 0.012480504 0.0181188639976218 0 +122 1 2.6941645 0.9987542 0.0017984251616327061 1 +123 1 1.4203832 0.96005714 0.058807822104565324 1 +125 0 -1.4110423 0.009791615 0.014195929527819996 0 +128 1 1.252279 0.93800443 0.09233335226231662 1 +129 0 -1.5439607 0.006811143 0.0098600193325827368 0 +131 0 -1.3816847 0.010607217 0.01538471954762626 0 +132 1 2.225845 0.9954915 0.0065190905678293599 1 +133 0 -1.4103005 0.00981144 0.014224812805453237 0 +137 0 -1.4452715 0.008918809 0.012924844355246622 0 +138 0 -1.3805954 0.010638738 0.015430682441826436 0 +141 0 -1.4420875 0.008996633 0.013038135642134178 0 +144 0 -1.4135613 0.0097245965 0.014098288900044151 0 +145 0 ? ? ? 0 +147 0 -1.3754209 0.010789749 0.015650905014368756 0 +150 0 -1.3678652 0.011014068 0.015978094989696334 0 +151 1 1.5347289 0.97052735 0.043159224086010806 1 +152 1 2.504544 0.9979019 0.0030300744424636324 1 +154 0 -1.4369092 0.009124645 0.013224506778708933 0 +156 0 -1.3866146 0.010465707 0.015178388707488839 0 +161 0 -1.3712488 0.010913048 0.015830738926007847 0 +164 0 ? ? ? 0 +167 1 2.3709106 0.9969716 0.0043756762449618574 1 +169 0 -1.4165915 0.009644579 0.01398171930433075 0 +171 0 -1.3817354 0.010605751 0.015382580668868604 0 +173 1 3.937659 0.99995935 5.8647275968961343E-05 1 +174 1 1.5719634 0.9733226 0.03901008486188097 1 +176 0 -1.3816847 0.010607217 0.01538471954762626 0 +177 1 1.9459012 0.9903058 0.014054033186831696 1 +179 1 0.9116728 0.8555586 0.22506146540399588 1 +180 0 -1.3678652 0.011014068 0.015978094989696334 0 +181 0 -1.4369092 0.009124645 0.013224506778708933 0 +183 1 2.3378227 0.9966837 0.0047923361356905954 1 +187 1 3.3537698 0.99979717 0.00029265816204806702 1 +188 1 2.3298023 0.9966099 0.004899151567833957 1 +189 0 -1.3222847 0.012468447 0.018101249837415477 0 +191 1 2.7861626 0.9990327 0.0013962286319306914 1 +192 0 -1.409861 0.009823204 0.014241953620204527 0 +196 0 1.8399341 0.98706394 6.272458398321854 1 +198 0 -1.4369092 0.009124645 0.013224506778708933 0 +199 0 -1.4141624 0.009708673 0.014075090368352271 0 +201 1 2.7072368 0.9987982 0.001734885803824911 1 +202 0 -1.3817354 0.010605751 0.015382580668868604 0 +204 0 -1.3817354 0.010605751 0.015382580668868604 0 +205 1 3.089224 0.99957985 0.0006062802271575527 1 +206 1 1.777327 0.9846676 0.022291308306016867 1 +207 0 -1.3678652 0.011014068 0.015978094989696334 0 +209 0 -1.3619052 0.01119426 0.016240976596509825 0 +210 1 3.5197217 0.99987155 0.00018532320875813371 1 +211 1 2.3830538 0.9970709 0.0042319866825224874 1 +212 0 -1.3817354 0.010605751 0.015382580668868604 0 +216 0 -1.4110423 0.009791615 0.014195929527819996 0 +218 1 2.3298595 0.99661046 0.0048983750135370039 1 +219 0 -1.2886547 0.013661583 0.019845368338199321 0 +223 1 1.633422 0.97738165 0.033006081115078849 1 +226 1 2.4771943 0.9977382 0.0032668086828737306 1 +228 0 -1.3678652 0.011014068 0.015978094989696334 0 +233 1 1.5488453 0.97161907 0.041537289441038114 1 +237 1 1.9858484 0.9913067 0.012596581290247206 1 +239 1 1.4558253 0.9636358 0.053440099405225483 1 +240 0 -1.2595147 0.014785988 0.021490948851795963 0 +241 0 -1.3571875 0.011338957 0.016452108687224885 0 +242 0 -1.3816847 0.010607217 0.01538471954762626 0 +244 0 -1.3817354 0.010605751 0.015382580668868604 0 +246 1 2.7713618 0.9989925 0.0014542440703349414 1 +247 1 0.91193277 0.855647 0.22491231777740134 1 +248 0 -1.350693 0.011541179 0.016747230629073448 0 +249 0 ? ? ? 0 +250 0 -1.4159538 0.009661366 0.014006174015630656 0 +252 0 1.1911402 0.9274629 3.7851366411407636 1 +254 1 2.0041015 0.99172926 0.011981773973740643 1 +257 0 -1.4141624 0.009708673 0.014075090368352271 0 +258 0 -1.3797933 0.01066201 0.01546461810748212 0 +259 0 1.1689192 0.9232376 3.7034567806518521 1 +260 1 2.1825306 0.9949233 0.0073427935751854403 1 +262 1 2.4699996 0.99769306 0.0033320531592707222 1 +267 1 1.2059503 0.93015885 0.10445097261233313 1 +268 1 2.6487386 0.99858844 0.0020378852032415661 1 +269 0 -1.3817354 0.010605751 0.015382580668868604 0 +271 0 -1.3707396 0.010928193 0.015852830032950474 0 +272 1 1.2059503 0.93015885 0.10445097261233313 1 +275 0 ? ? ? 0 +276 0 -1.4141624 0.009708673 0.014075090368352271 0 +277 0 -1.4110423 0.009791615 0.014195929527819996 0 +278 0 -1.3817354 0.010605751 0.015382580668868604 0 +279 1 1.879127 0.98837185 0.016874174669693685 1 +280 0 -1.3797933 0.01066201 0.01546461810748212 0 +283 1 1.4969043 0.96739787 0.047818735639454678 1 +284 1 1.5926834 0.9747646 0.036874257920139604 1 +285 1 3.603128 0.9998979 0.00014731066100425312 1 +288 1 1.0253819 0.89011955 0.16792897636163284 1 +290 0 -1.4369092 0.009124645 0.013224506778708933 0 +291 0 -1.3817354 0.010605751 0.015382580668868604 0 +293 1 1.1632752 0.9221291 0.11695935822097531 1 +296 0 0.92341477 0.8595081 2.8314411109622024 1 +297 0 ? ? ? 0 +299 1 1.8574562 0.9876657 0.017905267696866289 1 +300 1 2.0965576 0.99357617 0.0092975240440090982 1 +301 0 -1.3817354 0.010605751 0.015382580668868604 0 +303 0 -1.3817354 0.010605751 0.015382580668868604 0 +304 1 1.3020439 0.94551027 0.080834969005106949 1 +308 1 2.2036934 0.9952093 0.0069281615165738579 1 +309 0 -1.0833744 0.023794902 0.034743809724878026 0 +311 0 -1.4369092 0.009124645 0.013224506778708933 0 +312 1 1.3475217 0.95161307 0.071553010338479897 1 +314 0 -1.3953571 0.010219334 0.014819233017852235 0 +316 1 1.21495 0.93175155 0.10198278213026167 1 +317 1 2.4406166 0.9974991 0.0036125432073197024 1 +319 0 1.2473496 0.93721044 3.9933315017654505 1 +321 0 ? ? ? 0 +323 1 1.3046718 0.94588184 0.080268116762996347 1 +327 0 -1.4110423 0.009791615 0.014195929527819996 0 +328 1 1.2367946 0.9354784 0.096223769120927241 1 +329 1 1.978216 0.99112374 0.012862914334868926 1 +331 0 -1.3299384 0.012211618 0.017726094534328357 0 +332 0 -1.3626817 0.011170618 0.016206483223402176 0 +333 1 1.3505416 0.95199454 0.07097479822361287 1 +336 1 1.3618311 0.9533953 0.068853570897936811 1 +338 0 -1.3953571 0.010219334 0.014819233017852235 0 +343 0 -1.4369092 0.009124645 0.013224506778708933 0 +344 1 2.7494042 0.9989298 0.0015448010197468586 1 +346 0 -1.406384 0.009916759 0.014378270512468868 0 +347 0 -1.4153636 0.009676928 0.01402884366582037 0 +348 1 0.6176726 0.72499835 0.46395037336430689 1 +349 1 1.287628 0.94342875 0.084014521384112653 1 +350 0 -1.4341567 0.009193423 0.013324649921067406 0 +352 0 0.9037058 0.85282654 2.7644105211579557 1 +353 1 2.3469493 0.99676573 0.0046736232544708298 1 +354 0 -1.4110423 0.009791615 0.014195929527819996 0 +355 0 -1.4190902 0.0095790895 0.013886320723743368 0 +358 1 1.8351793 0.98689574 0.019030414748242817 1 +360 1 4.061983 0.99997115 4.1620401857849153E-05 1 +361 1 1.7376808 0.98292935 0.024840372715091329 1 +366 1 3.5150821 0.9998699 0.00018773127719125943 1 +368 0 -1.3847388 0.0105193285 0.015256568961339506 0 +370 0 -1.3891975 0.010392313 0.015071387913992663 0 +371 0 -1.3847388 0.0105193285 0.015256568961339506 0 +373 0 -1.4033726 0.0099984985 0.01449738158669686 0 +376 0 -1.4110423 0.009791615 0.014195929527819996 0 +377 0 -1.3953571 0.010219334 0.014819233017852235 0 +378 0 -1.4040215 0.00998083 0.01447163467375785 0 +379 0 -1.2613459 0.014712721 0.021383664324646709 0 +381 1 2.365325 0.9969248 0.0044433860714477719 1 +383 0 -1.4420875 0.008996633 0.013038135642134178 0 +384 0 -1.4420875 0.008996633 0.013038135642134178 0 +387 0 -1.2890881 0.013645511 0.019821860619724973 0 +388 0 -1.4101909 0.00981437 0.014229083063797844 0 +389 0 -1.3600347 0.0112514105 0.016324363333466204 0 +391 1 2.3726792 0.99698627 0.0043544582781243173 1 +392 0 -1.4141624 0.009708673 0.014075090368352271 0 +395 0 -1.4141624 0.009708673 0.014075090368352271 0 +396 0 -1.3797933 0.01066201 0.01546461810748212 0 +398 0 -1.412035 0.009765151 0.014157372352491168 0 +399 0 -1.3883209 0.010417163 0.015107616527867391 0 +404 0 -1.4320167 0.009247248 0.013403025423618877 0 +406 0 -1.38088 0.010630494 0.015418660922384425 0 +409 0 -1.3833861 0.010558165 0.015313194445956742 0 +413 0 -1.3351166 0.012040828 0.017476672593484927 0 +414 1 1.7379416 0.9829414 0.02482270090442612 1 +415 0 -1.177919 0.018441886 0.02685440822869898 0 +416 1 2.475959 0.9977305 0.0032779267536443934 1 +418 0 -1.2199697 0.01645883 0.023942651222751188 0 +419 0 -1.4888638 0.00791806 0.011468811169594837 0 +422 0 -1.2906637 0.013587246 0.019736641029675581 0 +423 0 -1.3411375 0.011845209 0.017191041799475098 0 +428 0 -1.4110423 0.009791615 0.014195929527819996 0 +429 0 -1.4135613 0.0097245965 0.014098288900044151 0 +430 0 -1.4150264 0.0096858265 0.014041807408869485 0 +434 0 1.4896405 0.9667612 4.910986668241053 1 +436 1 1.4430777 0.96238565 0.055312957370120222 1 +439 0 -1.4459052 0.008903398 0.012902411666770523 0 +440 1 2.32509 0.99656576 0.0049630893056075631 1 +441 0 -1.2318493 0.015937632 0.023178341085675704 0 +442 0 -1.332229 0.012135775 0.017615327502906994 0 +449 1 2.659151 0.99862826 0.0019803629470412963 1 +450 0 -1.3900899 0.010367074 0.015034594053130974 0 +451 0 -1.4459052 0.008903398 0.012902411666770523 0 +452 0 -1.435998 0.009147358 0.013257576977546382 0 +453 1 2.367775 0.99694544 0.0044135416034077829 1 +454 0 -1.4639792 0.008474834 0.012278705163694668 0 +455 1 0.40747887 0.59644 0.74555103861928662 1 +456 1 2.4565005 0.99760586 0.0034581548754549166 1 +457 1 2.6429174 0.9985657 0.0020707806979284151 1 +464 0 -1.4465233 0.008888394 0.012880570396540716 0 +465 1 2.6142442 0.998448 0.0022407814121053644 1 +466 1 2.4056876 0.99724734 0.0039767271978899798 1 +467 1 2.0223577 0.99213153 0.011396697171978555 1 +474 0 -1.4459052 0.008903398 0.012902411666770523 0 +480 0 -1.433779 0.009202899 0.013338448116221426 0 +482 1 4.237635 0.9999822 2.571163541833834E-05 1 +483 1 2.7514598 0.9989358 0.0015361066172870541 1 +484 0 -1.4140018 0.009712924 0.014081284114583212 0 +487 1 3.2816238 0.9997526 0.00035699415723449123 1 +489 1 0.4076561 0.59655744 0.74526704319104475 1 +492 0 -1.4012095 0.010057626 0.014583549074493642 0 +493 1 2.5301332 0.9980444 0.0028241377703653479 1 +495 0 -1.4011749 0.010058574 0.014584930771258478 0 +497 0 -1.4488631 0.008831819 0.012798220217601158 0 +501 0 -1.4129081 0.009741933 0.014123546118409658 0 +502 0 -1.4070108 0.009899829 0.014353600507544561 0 +504 0 -1.4369092 0.009124645 0.013224506778708933 0 +507 0 -1.2911807 0.01356818 0.019708755938698461 0 +510 0 -1.4369092 0.009124645 0.013224506778708933 0 +513 0 -1.4011749 0.010058574 0.014584930771258478 0 +514 1 2.5227704 0.9980044 0.0028819521693530998 1 +517 0 -1.3953571 0.010219334 0.014819233017852235 0 +519 1 1.9664453 0.990834 0.013284723366218894 1 +520 0 -1.4535215 0.008720246 0.012635830063894472 0 +521 0 -1.4590681 0.00858922 0.012445149341715065 0 +522 1 1.5380176 0.97078526 0.042775891216072504 1 +523 1 1.7925903 0.98528934 0.021380653327047233 1 +527 0 -1.3761241 0.010769103 0.015620793844278126 0 +528 0 -1.3683099 0.011000739 0.01595865251196367 0 +529 0 -1.4012095 0.010057626 0.014583549074493642 0 +531 0 -1.38088 0.010630494 0.015418660922384425 0 +532 0 -1.3678652 0.011014068 0.015978094989696334 0 +533 0 -1.4141624 0.009708673 0.014075090368352271 0 +534 0 -1.4135613 0.0097245965 0.014098288900044151 0 +535 0 -1.3155776 0.0126978895 0.018436483688407564 0 +538 0 -1.4129081 0.009741933 0.014123546118409658 0 +539 0 -1.4050834 0.00995198 0.014429593182483781 0 +540 0 -1.3563625 0.011364451 0.016489311100219266 0 +541 0 -1.4452715 0.008918809 0.012924844355246622 0 +544 0 -1.3483434 0.011615216 0.016855293780747255 0 +546 1 2.72134 0.9988439 0.0016688526073658334 1 +547 0 -1.4259484 0.009401588 0.013627786320506213 0 +548 0 -1.4305096 0.009285345 0.01345850137331395 0 +549 1 1.7659012 0.98418534 0.022998070923809989 1 +557 0 -1.4341567 0.009193423 0.013324649921067406 0 +558 0 -1.4135613 0.0097245965 0.014098288900044151 0 +559 0 -1.409861 0.009823204 0.014241953620204527 0 +560 0 -1.3707396 0.010928193 0.015852830032950474 0 +561 0 -1.3707396 0.010928193 0.015852830032950474 0 +563 0 -1.4141624 0.009708673 0.014075090368352271 0 +565 1 3.0260675 0.9995001 0.00072138957424316519 1 +566 0 -1.347076 0.011655343 0.016913866260533591 0 +569 1 2.7834864 0.9990255 0.0014065576193970416 1 +577 0 -1.4110423 0.009791615 0.014195929527819996 0 +578 0 -1.4110423 0.009791615 0.014195929527819996 0 +581 1 2.2628958 0.995927 0.0058881306368555441 1 +582 1 2.0087006 0.9918325 0.011831602721986766 1 +584 0 -1.4244834 0.009439227 0.013682604912380625 0 +586 1 3.4116895 0.999827 0.00024956841128260089 1 +590 1 1.685864 0.980363 0.028612041867767141 1 +593 0 -1.4140018 0.009712924 0.014081284114583212 0 +594 1 1.5861186 0.9743161 0.03753815936053715 1 +600 0 -1.4141624 0.009708673 0.014075090368352271 0 +602 0 -1.4129081 0.009741933 0.014123546118409658 0 +604 1 1.673604 0.9797025 0.029584413364803362 1 +606 0 -1.4413451 0.009014877 0.013064694891877983 0 +607 0 -1.4369092 0.009124645 0.013224506778708933 0 +609 0 -1.4459052 0.008903398 0.012902411666770523 0 +612 1 4.2357197 0.9999821 2.5797628278879132E-05 1 +613 0 -1.4190167 0.009581008 0.013889115342110844 0 +614 0 -1.3992448 0.010111629 0.014662252251201175 0 +617 0 ? ? ? 0 +618 0 -1.4129081 0.009741933 0.014123546118409658 0 +619 0 -1.409861 0.009823204 0.014241953620204527 0 +621 0 0.664872 0.75013816 2.0007975319694276 1 +622 0 -1.3458307 0.011694909 0.01697162269085075 0 +624 0 -1.3933338 0.010275838 0.014901595467036012 0 +627 0 -1.2435194 0.015441447 0.022451086588681352 0 +629 0 -1.4465233 0.008888394 0.012880570396540716 0 +633 1 1.433203 0.96138895 0.056807879979561222 1 +634 0 -1.4452715 0.008918809 0.012924844355246622 0 +638 0 -1.4465233 0.008888394 0.012880570396540716 0 +639 0 -1.4341567 0.009193423 0.013324649921067406 0 +641 0 -1.4141624 0.009708673 0.014075090368352271 0 +642 0 -1.4141624 0.009708673 0.014075090368352271 0 +644 0 -1.4420875 0.008996633 0.013038135642134178 0 +645 0 -1.4141624 0.009708673 0.014075090368352271 0 +649 0 -1.4141624 0.009708673 0.014075090368352271 0 +652 0 -1.4050529 0.009952808 0.014430799663108925 0 +653 0 -1.4129081 0.009741933 0.014123546118409658 0 +654 0 -1.3797933 0.01066201 0.01546461810748212 0 +656 0 -1.409861 0.009823204 0.014241953620204527 0 +657 0 0.7602914 0.79609644 2.2940411601833461 1 +660 0 -1.4110423 0.009791615 0.014195929527819996 0 +661 0 -1.3761241 0.010769103 0.015620793844278126 0 +665 0 -1.4369092 0.009124645 0.013224506778708933 0 +668 1 1.4711138 0.9650824 0.051275957593795329 1 +670 1 1.8996595 0.98900396 0.015951802670619887 1 +678 0 -1.4369092 0.009124645 0.013224506778708933 0 +679 0 -1.4420875 0.008996633 0.013038135642134178 0 +680 1 4.0464444 0.9999699 4.3426272919049767E-05 1 +681 1 2.396061 0.99717367 0.0040833097880269963 1 +682 0 -1.3654106 0.01108793 0.016085846470416602 0 +683 0 -1.4369092 0.009124645 0.013224506778708933 0 +685 0 -1.4369092 0.009124645 0.013224506778708933 0 +688 0 -1.4465233 0.008888394 0.012880570396540716 0 +689 0 -1.1964769 0.017539352 0.025528473093602299 0 +691 1 1.5043631 0.96803945 0.046862248285937283 1 +692 0 -1.4452715 0.008918809 0.012924844355246622 0 +693 0 -1.3783048 0.010705326 0.015527784464396777 0 +694 0 -1.4078116 0.009878239 0.014322141740327976 0 +696 1 2.3273046 0.99658656 0.0049329752272506688 1 +697 1 1.7718921 0.9844401 0.022624686610027662 1 +698 1 2.0078473 0.9918135 0.01185926010959627 1 +0 0 -1.1421579 0.006991991 0.010122740954419757 0 +1 0 0.9726735 0.9051856 3.3987497032818106 1 +2 0 -1.1365752 0.007125426 0.010316615822619619 0 +3 0 0.9583137 0.90089846 3.3349486633777174 1 +4 0 -1.1396834 0.0070508276 0.010208224686015586 0 +7 0 -1.1484356 0.006844906 0.0099090642676198451 0 +12 1 0.84079856 0.85893667 0.21937633481036786 1 +13 0 -1.1307771 0.007266687 0.010521889739744694 0 +14 1 1.2031609 0.9544467 0.06726350090433679 1 +15 1 0.975671 0.90605927 0.14232267516500977 1 +16 0 -1.1400965 0.0070409696 0.010193901608002667 0 +17 0 -1.1443123 0.0069411644 0.010048899444453792 0 +19 0 -1.1407655 0.0070250365 0.01017075224226344 0 +22 0 -1.1431679 0.0069681183 0.010088057996026958 0 +23 1 ? ? ? 0 +24 0 -1.1549612 0.006695268 0.0096917108714694045 0 +26 0 -1.1140122 0.0076909643 0.011138604428553758 0 +27 0 -1.1377332 0.007097542 0.010276099200573898 0 +29 0 -1.1423135 0.0069883084 0.010117390905200944 0 +30 0 -1.1255252 0.007397038 0.01071133503755105 0 +33 0 -1.147387 0.0068692584 0.0099444396705657186 0 +34 0 -1.1410553 0.007018146 0.010160741217356693 0 +36 1 1.3150669 0.96844256 0.046261613383462345 1 +38 1 1.075442 0.9312908 0.10269636006637334 1 +39 1 0.87801623 0.87362635 0.19491172169050572 1 +42 1 1.1366 0.9434947 0.083913715591640861 1 +43 1 -1.2082165 0.0055895424 7.483054094552843 0 +47 0 -1.1512741 0.006779413 0.0098139298690182678 0 +49 1 1.0526962 0.92615813 0.1106695578447357 1 +53 1 1.1666713 0.94871974 0.07594612940253459 1 +55 1 1.1210563 0.9406013 0.088344785203172155 1 +57 1 1.0027217 0.9136226 0.13032972825969921 1 +58 1 0.9314057 0.892399 0.16423917686733305 1 +59 1 1.0253713 0.9195267 0.12103663452565497 1 +61 0 -1.1388069 0.0070717838 0.010238673031863122 0 +62 1 1.1175295 0.93992573 0.089381330505907935 1 +65 1 1.1504183 0.9459551 0.080156391130935659 1 +67 1 0.99326116 0.9110423 0.13441009717524816 1 +75 0 -1.1495842 0.0068183294 0.0098704584735539429 0 +78 0 -1.1263603 0.0073761577 0.010680987032854883 0 +80 0 -1.1358691 0.007142482 0.0103413989074172 0 +81 0 -1.1374917 0.0071033486 0.010284536545637598 0 +83 0 -1.1541876 0.006712838 0.0097172299110602385 0 +84 1 1.1617104 0.9478904 0.077207836648080577 1 +85 1 1.0497433 0.9254665 0.11174736096139688 1 +86 1 0.9147182 0.88681114 0.17330120618704528 1 +87 1 1.0544082 0.9265564 0.11004928523023362 1 +89 0 -1.1611012 0.0065574422 0.0094915436599937088 0 +94 0 -1.1507537 0.006791373 0.009831301769275139 0 +101 1 0.7321916 0.8078508 0.30783926425261254 1 +103 1 -1.2068802 0.0056149308 7.4765160514003322 0 +107 1 1.0343939 0.9217744 0.11751441372031256 1 +110 0 -1.1421878 0.0069912826 0.010121711941135984 0 +114 0 -1.1024808 0.007996975 0.011583575580737138 0 +116 0 0.8794645 0.8741706 2.9904590637227133 1 +118 0 -1.1577381 0.0066325828 0.0096006681675805308 0 +119 0 -1.1250697 0.007408452 0.010727925189033533 0 +124 1 1.1666223 0.9487116 0.075958547044740976 1 +126 1 1.17557 0.9501762 0.073733055555185265 1 +127 0 -1.1471906 0.00687383 0.0099510811109006243 0 +130 0 -1.1348375 0.0071674744 0.010377715316368026 0 +134 0 -1.1607496 0.00656526 0.009502896449406141 0 +135 0 -1.1452248 0.0069197463 0.010017784026900324 0 +136 0 -1.1400965 0.0070409696 0.010193901608002667 0 +139 0 ? ? ? 0 +140 0 -1.139174 0.007063 0.010225910534514665 0 +142 1 1.1103708 0.93853235 0.091521617373989284 1 +143 0 -1.0984974 0.008105465 0.011741363615556305 0 +146 1 0.8372512 0.8574645 0.2218511644917357 1 +148 0 -1.2880107 0.0042635775 0.0061641922538819361 0 +149 1 1.2488376 0.96075964 0.057752548308823239 1 +153 0 -1.1157168 0.0076467264 0.011074289520616544 0 +155 1 0.9783753 0.90684134 0.14107793814057104 1 +157 0 -1.1507537 0.006791373 0.009831301769275139 0 +158 0 ? ? ? 0 +159 1 1.2657585 0.9628782 0.054574739263364667 1 +160 1 1.2091928 0.9553327 0.06592485292809229 1 +162 0 -1.1471906 0.00687383 0.0099510811109006243 0 +163 0 -1.1461331 0.0068984935 0.0099869094286237654 0 +165 0 -1.1402277 0.0070378436 0.010189359794395579 0 +166 1 1.106156 0.9376979 0.092804912579276269 1 +168 0 -1.1471906 0.00687383 0.0099510811109006243 0 +170 0 -1.139174 0.007063 0.010225910534514665 0 +172 0 -1.1512741 0.006779413 0.0098139298690182678 0 +175 1 1.107263 0.93791807 0.092466195096246012 1 +178 0 -1.1443123 0.0069411644 0.010048899444453792 0 +182 0 -1.1407655 0.0070250365 0.01017075224226344 0 +184 1 1.1061851 0.93770367 0.092796017248526444 1 +185 0 -1.1329874 0.007212513 0.010443163080674363 0 +186 1 1.0154562 0.9169888 0.1250239975549377 1 +190 1 1.283762 0.96501154 0.051381904442960757 1 +193 0 -1.1549612 0.006695268 0.0096917108714694045 0 +194 0 -1.1471906 0.00687383 0.0099510811109006243 0 +195 0 -1.1443123 0.0069411644 0.010048899444453792 0 +197 0 -1.1371826 0.0071107876 0.010295345486010998 0 +200 1 1.2376518 0.9592959 0.059952161093617132 1 +203 0 -1.1421579 0.006991991 0.010122740954419757 0 +208 0 -1.1369494 0.007116404 0.010303506205067439 0 +213 1 1.3092372 0.9678293 0.047175498258298046 1 +214 1 1.3174671 0.96869177 0.045890415858460022 1 +215 1 1.1947613 0.95318484 0.069172083991982519 1 +217 0 -1.1549612 0.006695268 0.0096917108714694045 0 +220 0 -1.151182 0.006781527 0.0098170000190954988 0 +221 1 1.2784731 0.9643974 0.052300377083517792 1 +222 1 -1.0862726 0.008447611 6.8872409211527614 0 +224 1 1.2527163 0.9612553 0.05700842935438296 1 +225 0 -1.1512741 0.006779413 0.0098139298690182678 0 +227 1 1.1860297 0.951838 0.071212017847282261 1 +229 1 1.3019414 0.9670455 0.048344344952710637 1 +230 1 1.1431115 0.9446669 0.082122352379010777 1 +231 1 1.1684545 0.9490148 0.075497534449766812 1 +232 0 0.9967845 0.9120113 3.5065383488798765 1 +234 0 -1.1663417 0.0064420407 0.0093239651597370656 0 +235 0 ? ? ? 0 +236 1 1.2541354 0.96143514 0.05673856185758594 1 +238 1 1.2263106 0.9577586 0.062266011536222791 1 +243 0 -1.1176138 0.007597792 0.011003149740191447 0 +245 0 -1.1319131 0.0072387946 0.010481355171756924 0 +251 1 1.2275534 0.95792973 0.062008264917063245 1 +253 1 1.1366 0.9434947 0.083913715591640861 1 +255 1 0.9725831 0.9051591 0.14375667243144519 1 +256 0 -1.139174 0.007063 0.010225910534514665 0 +261 1 1.2392927 0.9595139 0.059624384704991769 1 +263 1 1.2355157 0.95901054 0.060381421460153704 1 +264 1 1.1225077 0.94087726 0.087921565028788271 1 +265 0 -1.1442158 0.006943432 0.010052194017572808 0 +266 1 1.1370882 0.94358337 0.08377810372062873 1 +270 1 1.1017067 0.93680555 0.094178477660174398 1 +273 1 0.8982152 0.8810377 0.18272432104119168 1 +274 0 -1.1523645 0.00675442 0.0097776269726953576 0 +281 0 -1.147387 0.0068692584 0.0099444396705657186 0 +282 1 0.95462376 0.89976925 0.15237303733617075 1 +286 1 1.3084354 0.96774405 0.047302558900165076 1 +287 0 -1.1607496 0.00656526 0.009502896449406141 0 +289 1 1.1160771 0.93964547 0.089811568227469565 1 +292 1 ? ? ? 0 +294 0 ? ? ? 0 +295 1 1.1387173 0.94387835 0.083327158339233473 1 +298 0 -1.1688498 0.006387524 0.0092448062126711424 0 +302 1 1.3397653 0.9709176 0.042579258921414516 1 +305 1 1.1629357 0.9480964 0.07689434709997052 1 +306 0 -1.1549612 0.006695268 0.0096917108714694045 0 +307 0 -1.1549612 0.006695268 0.0096917108714694045 0 +310 0 -1.164732 0.0064772717 0.0093751232145894009 0 +313 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +315 0 ? ? ? 0 +318 0 -1.2086128 0.0055820365 0.0080757368621154728 0 +320 1 1.1274756 0.9418127 0.086487926983042085 1 +322 0 -1.1471906 0.00687383 0.0099510811109006243 0 +324 0 -1.1549612 0.006695268 0.0096917108714694045 0 +325 0 -1.1400728 0.0070415353 0.01019472364193548 0 +326 1 1.1130694 0.9390611 0.090709056949963965 1 +330 1 1.1741606 0.94994813 0.074079351648233677 1 +334 1 1.0892959 0.9342531 0.098114654462840872 1 +335 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +337 0 -1.1549612 0.006695268 0.0096917108714694045 0 +339 1 1.0955105 0.935543 0.096124128788320856 1 +340 1 1.1813669 0.9511038 0.072325284880902799 1 +341 0 -1.1549612 0.006695268 0.0096917108714694045 0 +342 0 -1.1430825 0.0069701327 0.010090984630184313 0 +345 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +351 0 -1.1507537 0.006791373 0.009831301769275139 0 +356 1 -1.1481237 0.0068521397 7.1892297156089855 0 +357 1 1.2755169 0.9640496 0.052820753741755856 1 +359 1 1.1539546 0.94656837 0.079221380502525668 1 +362 0 -1.1425558 0.006982575 0.010109061429518538 0 +363 0 -1.084903 0.008486823 0.012296149454680742 0 +364 0 -1.1507537 0.006791373 0.009831301769275139 0 +365 0 -1.1469072 0.00688043 0.0099606685689672403 0 +367 1 1.2065824 0.9549513 0.066500954181848593 1 +369 0 -1.1282009 0.0073303413 0.010614398275747035 0 +372 0 -1.1382743 0.00708455 0.010257221919190783 0 +374 0 -1.1410553 0.007018146 0.010160741217356693 0 +375 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +380 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +382 0 -1.1365077 0.0071270545 0.010318981994411457 0 +385 0 -1.1479244 0.006856767 0.0099262939070172643 0 +386 1 1.1005563 0.9365729 0.09453678662211823 1 +390 0 -1.1473866 0.006869269 0.0099444552290070438 0 +393 0 -1.1293422 0.0073020733 0.010573315648837111 0 +394 0 -1.134171 0.007183667 0.010401244907416443 0 +397 0 -1.1359171 0.007141323 0.010339714751142249 0 +400 1 1.1509389 0.9460458 0.080018041517608593 1 +401 0 -1.139174 0.007063 0.010225910534514665 0 +402 0 -1.1130127 0.007717023 0.011176491313271456 0 +403 0 -1.1068444 0.007879782 0.011413147589468792 0 +405 0 -1.1512741 0.006779413 0.0098139298690182678 0 +407 0 -1.1512741 0.006779413 0.0098139298690182678 0 +408 0 -1.1115428 0.007755502 0.011232437575084145 0 +410 0 -1.1512741 0.006779413 0.0098139298690182678 0 +411 0 ? ? ? 0 +412 1 1.2862241 0.9652939 0.050959856297659256 1 +417 0 -1.1512741 0.006779413 0.0098139298690182678 0 +420 0 -1.110955 0.007770944 0.011254889675297716 0 +421 1 1.2906268 0.96579343 0.050213444357725026 1 +424 0 -1.139174 0.007063 0.010225910534514665 0 +425 1 1.3378967 0.9707371 0.042847464938219351 1 +426 0 -1.1053871 0.00791873 0.011469786293421675 0 +427 1 1.1061093 0.9376886 0.092819218587062283 1 +431 0 -1.1776268 0.006200332 0.0089730352447954451 0 +432 0 -1.1432939 0.0069651455 0.010083739090445843 0 +433 0 -1.1199065 0.007539063 0.010917775938699774 0 +435 1 1.1695958 0.9492028 0.075211775169929426 1 +437 0 -1.1359171 0.007141323 0.010339714751142249 0 +438 0 -1.1278342 0.0073394463 0.010627631184494359 0 +443 0 -1.1349095 0.007165727 0.010375176500964851 0 +444 0 -1.1466866 0.006885574 0.0099681414593567122 0 +445 0 -1.1430825 0.0069701327 0.010090984630184313 0 +446 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +447 0 -1.1333517 0.0072036223 0.010430243137407541 0 +448 0 -1.1293422 0.0073020733 0.010573315648837111 0 +458 0 -1.1317519 0.0072427453 0.010487096354881752 0 +459 0 -1.1302816 0.0072788885 0.01053962139132392 0 +460 0 -1.1291116 0.007307775 0.010581601786164479 0 +461 0 -1.0969267 0.008148644 0.011804168027659273 0 +462 0 -1.1282747 0.0073285103 0.010611737225750506 0 +463 0 -1.1412017 0.0070146676 0.010155687357411498 0 +468 0 -1.1359171 0.007141323 0.010339714751142249 0 +469 0 -1.1486369 0.0068402397 0.0099022857045564829 0 +470 0 -1.1255252 0.007397038 0.01071133503755105 0 +471 0 -1.1282747 0.0073285103 0.010611737225750506 0 +472 0 -1.1144809 0.0076787774 0.011120886400783731 0 +473 0 -1.1359171 0.007141323 0.010339714751142249 0 +475 0 -1.139174 0.007063 0.010225910534514665 0 +476 0 -1.1340168 0.0071874196 0.010406698186655973 0 +477 0 -1.1359171 0.007141323 0.010339714751142249 0 +478 0 -1.1221378 0.0074823424 0.010835325857362313 0 +479 1 1.1185335 0.9401188 0.089085033369214639 1 +481 0 -1.1052345 0.007922819 0.011475731854348652 0 +485 0 -1.0989596 0.008092803 0.011722946672065039 0 +486 0 -1.1255252 0.007397038 0.01071133503755105 0 +488 1 0.88935137 0.87783283 0.18798186852256374 1 +490 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +491 1 1.0793833 0.9321459 0.10137232084148741 1 +494 0 0.91586417 0.8872028 3.1481968304796748 1 +496 0 -1.1293422 0.0073020733 0.010573315648837111 0 +498 0 -1.1400965 0.0070409696 0.010193901608002667 0 +499 0 -1.1400965 0.0070409696 0.010193901608002667 0 +500 0 -1.1407655 0.0070250365 0.01017075224226344 0 +503 0 -1.1443123 0.0069411644 0.010048899444453792 0 +505 0 -1.1157738 0.0076452517 0.011072145514068717 0 +506 1 1.2465644 0.9604663 0.058193062150963197 1 +508 0 -1.1333517 0.0072036223 0.010430243137407541 0 +509 0 -1.1430825 0.0069701327 0.010090984630184313 0 +511 0 -1.1377332 0.007097542 0.010276099200573898 0 +512 0 -1.1333517 0.0072036223 0.010430243137407541 0 +515 1 1.107773 0.9380193 0.092310525429242715 1 +516 0 -1.1293422 0.0073020733 0.010573315648837111 0 +518 0 -1.1339877 0.007188128 0.01040772740322852 0 +524 0 -1.1431679 0.0069681183 0.010088057996026958 0 +525 0 -1.1359911 0.0071395333 0.010337114428541148 0 +526 0 -1.1359171 0.007141323 0.010339714751142249 0 +530 1 1.114476 0.93933505 0.090288255263633016 1 +536 0 -1.1421579 0.006991991 0.010122740954419757 0 +537 0 -1.1407323 0.0070258277 0.010171901718313003 0 +542 0 -1.1109444 0.007771223 0.011255295239501693 0 +543 0 -1.1400965 0.0070409696 0.010193901608002667 0 +545 0 -1.1377332 0.007097542 0.010276099200573898 0 +550 0 -1.1431679 0.0069681183 0.010088057996026958 0 +551 0 -1.1549612 0.006695268 0.0096917108714694045 0 +552 0 -1.1409097 0.0070216088 0.010165772092042155 0 +553 0 -1.0788838 0.008661311 0.012550059383235973 0 +554 0 -1.139174 0.007063 0.010225910534514665 0 +555 0 -1.1666757 0.006434754 0.0093133845983447393 0 +556 0 -1.1066437 0.007885136 0.011420933400330548 0 +562 0 -1.1549612 0.006695268 0.0096917108714694045 0 +564 0 -1.1497022 0.0068156044 0.0098665000733611837 0 +567 0 -1.120484 0.0075243427 0.010896377602635154 0 +568 1 1.0339276 0.9216596 0.11769409934522305 1 +570 1 1.1550791 0.94676197 0.078926345007957768 1 +571 1 1.2442987 0.9601719 0.058635412147082483 1 +572 0 -1.1431679 0.0069681183 0.010088057996026958 0 +573 0 -1.1512741 0.006779413 0.0098139298690182678 0 +574 1 1.1978174 0.9536478 0.068471556189648647 1 +575 0 -1.1407323 0.0070258277 0.010171901718313003 0 +576 0 -1.1377332 0.007097542 0.010276099200573898 0 +579 0 -1.1549612 0.006695268 0.0096917108714694045 0 +580 0 -1.1362189 0.0071340282 0.010329115234978077 0 +583 0 -1.139174 0.007063 0.010225910534514665 0 +585 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +587 0 -1.1432939 0.0069651455 0.010083739090445843 0 +588 1 1.1637968 0.9482407 0.076674781707131864 1 +589 0 -1.1333517 0.0072036223 0.010430243137407541 0 +591 1 1.0249119 0.91941065 0.12121872350389055 1 +592 1 1.1490365 0.9457137 0.080524600270426447 1 +595 0 -1.1377332 0.007097542 0.010276099200573898 0 +596 0 -1.1382743 0.00708455 0.010257221919190783 0 +597 0 -1.1394427 0.0070565767 0.010216577738017128 0 +598 0 -1.1431679 0.0069681183 0.010088057996026958 0 +599 0 -1.101245 0.008030476 0.01163229717799592 0 +601 0 -1.1331393 0.007208805 0.01043777462538828 0 +603 1 1.1223485 0.94084704 0.087967902956234717 1 +605 1 1.2861114 0.965281 0.050979098365244051 1 +608 1 1.2308334 0.9583782 0.061333009405650819 1 +610 1 1.2462684 0.960428 0.058250631608633467 1 +611 1 1.1664525 0.94868344 0.076001329813000423 1 +615 0 -1.1365769 0.007125386 0.010316557632570115 0 +616 0 -1.1431679 0.0069681183 0.010088057996026958 0 +620 0 -1.1431679 0.0069681183 0.010088057996026958 0 +623 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +625 0 -1.1214182 0.0075005875 0.010861846615771373 0 +626 1 1.064452 0.9288535 0.10647700575935067 1 +628 0 -1.1430825 0.0069701327 0.010090984630184313 0 +630 0 -1.1183742 0.0075782635 0.01097476074389735 0 +631 0 -1.1377332 0.007097542 0.010276099200573898 0 +632 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +635 0 -1.1329685 0.0072129727 0.010443830971736847 0 +636 1 1.3563839 0.9724759 0.0402656176234173 1 +637 0 -1.1128399 0.0077215354 0.011183051767306329 0 +640 0 -1.1307266 0.0072679305 0.010523696596031364 0 +643 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +646 0 -1.1328661 0.0072154743 0.010447466145129364 0 +647 0 -1.1398593 0.007046628 0.01020212264498309 0 +648 1 1.3318609 0.9701467 0.043725151791969848 1 +650 0 -1.1218915 0.0074885823 0.010844395968216777 0 +651 0 -1.1318827 0.0072395396 0.010482437901385536 0 +655 0 -1.1431679 0.0069681183 0.010088057996026958 0 +658 1 1.1772354 0.9504444 0.073325861219672731 1 +659 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +662 0 -1.1460992 0.0068992847 0.009988058758204724 0 +663 0 -1.1460992 0.0068992847 0.009988058758204724 0 +664 0 -1.1373322 0.007107186 0.010290112527864072 0 +666 0 -1.1260815 0.007383123 0.010691110633935377 0 +667 0 -1.1471906 0.00687383 0.0099510811109006243 0 +669 1 1.1582375 0.9473023 0.078103235245799443 1 +671 0 -1.1328036 0.007217003 0.01044968771967735 0 +672 0 -1.1507537 0.006791373 0.009831301769275139 0 +673 0 -1.1263967 0.00737525 0.010679667951409594 0 +674 0 -1.1512741 0.006779413 0.0098139298690182678 0 +675 0 -1.130584 0.00727144 0.010528796409353352 0 +676 0 -1.1486369 0.0068402397 0.0099022857045564829 0 +677 0 -1.1333517 0.0072036223 0.010430243137407541 0 +684 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +686 0 -1.1476012 0.0068642762 0.0099372023077261389 0 +687 0 -1.1534802 0.0067289425 0.009740620900575073 0 +690 0 -1.1398593 0.007046628 0.01020212264498309 0 +695 0 -1.1430825 0.0069701327 0.010090984630184313 0 diff --git a/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer-out.txt b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer-out.txt new file mode 100644 index 0000000000..1adb36c592 --- /dev/null +++ b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer-out.txt @@ -0,0 +1,56 @@ +maml.exe CV tr=LdSvm{iter=1000 bias=-} threads=- dout=%Output% data=%Data% seed=1 +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Warning: Skipped 8 rows with missing feature/label values +Training calibrator. +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Warning: Skipped 8 rows with missing feature/label values +Training calibrator. +Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3785 (134.0/(134.0+220.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 124 | 10 | 0.9254 + negative || 12 | 208 | 0.9455 + ||====================== +Precision || 0.9118 | 0.9541 | +OVERALL 0/1 ACCURACY: 0.937853 +LOG LOSS/instance: 0.277794 +Test-set entropy (prior Log-Loss/instance): 0.956998 +LOG-LOSS REDUCTION (RIG): 0.709724 +AUC: 0.980495 +Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3191 (105.0/(105.0+224.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 100 | 5 | 0.9524 + negative || 18 | 206 | 0.9196 + ||====================== +Precision || 0.8475 | 0.9763 | +OVERALL 0/1 ACCURACY: 0.930091 +LOG LOSS/instance: 0.265110 +Test-set entropy (prior Log-Loss/instance): 0.903454 +LOG-LOSS REDUCTION (RIG): 0.706560 +AUC: 0.971641 + +OVERALL RESULTS +--------------------------------------- +AUC: 0.976068 (0.0044) +Accuracy: 0.933972 (0.0039) +Positive precision: 0.879611 (0.0322) +Positive recall: 0.938877 (0.0135) +Negative precision: 0.965216 (0.0111) +Negative recall: 0.932549 (0.0129) +Log-loss: 0.271452 (0.0063) +Log-loss reduction: 0.708142 (0.0016) +F1 Score: 0.907690 (0.0108) +AUPRC: 0.966424 (0.0047) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer-rp.txt b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer-rp.txt new file mode 100644 index 0000000000..c198b0f299 --- /dev/null +++ b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer-rp.txt @@ -0,0 +1,4 @@ +LdSvm +AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC /bias /iter Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.976068 0.933972 0.879611 0.938877 0.965216 0.932549 0.271452 0.708142 0.90769 0.966424 - 1000 LdSvm %Data% %Output% 99 0 0 maml.exe CV tr=LdSvm{iter=1000 bias=-} threads=- dout=%Output% data=%Data% seed=1 /bias:-;/iter:1000 + diff --git a/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer.txt b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer.txt new file mode 100644 index 0000000000..3d802aefa4 --- /dev/null +++ b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-CV-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +5 1 3.9617066 0.99747086 0.0036534058658982058 1 +6 0 1.8168322 0.8994931 3.3146335211879085 1 +8 0 -1.6938565 0.017901568 0.026060466871379716 0 +9 0 -1.110207 0.048584107 0.071851970244810354 0 +10 0 -1.0310725 0.055462856 0.082320562822649976 0 +11 0 -1.2838186 0.03622594 0.053233125000715482 0 +18 1 2.5304081 0.9692632 0.045039623268477103 1 +20 1 1.6496639 0.86950064 0.20174101034003178 1 +21 1 1.8026577 0.89720863 0.15648459584437696 1 +25 1 2.1921978 0.9455318 0.080802137510512365 1 +28 0 -1.2838186 0.03622594 0.053233125000715482 0 +31 0 -1.5638075 0.022416964 0.032708844858653462 0 +32 1 0.04440026 0.28155547 1.8285088958574633 1 +35 0 -1.2838186 0.03622594 0.053233125000715482 0 +37 0 -2.2645154 0.006613392 0.0095727970580140424 0 +40 0 ? ? ? 0 +41 1 -0.19681369 0.20383355 2.2945365618095623 0 +44 1 1.226375 0.7594088 0.39705143114780073 1 +45 0 -1.3092242 0.034692515 0.050939528801904527 0 +46 1 4.547542 0.9990992 0.0013001725901859199 1 +48 0 -1.8601606 0.013408965 0.019475917672364153 0 +50 1 0.24273674 0.35739407 1.4844124002681229 1 +51 1 1.1279857 0.72627056 0.46142100157693239 1 +52 1 1.6171751 0.86285496 0.21281002628898907 1 +54 1 -0.06805999 0.24319303 2.0398262115538679 0 +56 1 4.23633 0.99844086 0.0022511164479464001 1 +60 1 1.913834 0.9139474 0.12981695257199879 1 +63 1 0.5027561 0.46810454 1.0950973335082912 1 +64 0 -1.1371732 0.046430733 0.068590355669037276 0 +66 0 -1.8191407 0.0144012915 0.02092772931478808 0 +68 1 1.8546224 0.9053645 0.14342933496516677 1 +69 0 -1.233573 0.039452832 0.058071635508138972 0 +70 0 -1.4396573 0.027756434 0.040610312397176378 0 +71 1 -0.01552603 0.2606622 1.9397467203607552 0 +72 0 -1.15682 0.044919334 0.066305507272326383 0 +73 1 2.8685744 0.9828413 0.024969593473354958 1 +74 1 0.42785037 0.43537232 1.1996783985040143 1 +76 0 -0.8721561 0.07212528 0.10799806404570846 0 +77 0 -0.76458126 0.085910596 0.12959281749605428 0 +79 0 -1.6685361 0.018704446 0.027240371262152979 0 +82 0 -1.4572215 0.026931977 0.039387433601929173 0 +88 0 -1.8191407 0.0144012915 0.02092772931478808 0 +90 0 -1.4220948 0.028605305 0.041870486871704173 0 +91 0 -1.1811787 0.043110523 0.063575795808411911 0 +92 0 -1.8191407 0.0144012915 0.02092772931478808 0 +93 0 -1.1371732 0.046430733 0.068590355669037276 0 +95 0 -1.4220948 0.028605305 0.041870486871704173 0 +96 0 -1.0460658 0.054092724 0.080229326579365073 0 +97 0 -1.9356104 0.011756743 0.017061889077286219 0 +98 1 0.44360772 0.44222102 1.1771605068344531 1 +99 1 3.3757439 0.9929181 0.010253409279617097 1 +100 1 -0.26988548 0.18370011 2.4445755692977276 0 +102 0 -1.7044595 0.017575497 0.025581550676448673 0 +104 1 -0.01071214 0.26230296 1.930693979775796 0 +105 1 -0.78686327 0.08287204 3.5929707434820215 0 +106 1 4.2329383 0.9984315 0.0022646382316379881 1 +108 0 -0.7508426 0.08783408 0.13263182913764879 0 +109 1 3.1514504 0.9895145 0.015207203400812518 1 +111 1 1.8093723 0.8982965 0.1547364216210384 1 +112 1 1.864046 0.90678 0.14117551647386983 1 +113 1 1.0107481 0.68327355 0.5494648066088641 1 +115 0 0.37494186 0.4125705 0.76751239255957682 1 +117 1 2.7996807 0.98066556 0.028166875513010267 1 +120 0 -1.3778462 0.030857215 0.045218859110108284 0 +121 0 -0.99584746 0.05881134 0.087444157404548681 0 +122 1 1.9245611 0.9154248 0.12748668056371995 1 +123 1 0.93247306 0.6526482 0.61562253296003511 1 +125 0 -1.1371732 0.046430733 0.068590355669037276 0 +128 1 2.4193385 0.96285367 0.054611534028127434 1 +129 0 -3.508995 0.0007396859 0.0010675360491821061 0 +131 0 -1.5638075 0.022416964 0.032708844858653462 0 +132 1 4.254703 0.9984905 0.0021793756009357 1 +133 0 -1.2129154 0.040858 0.060183675071907958 0 +137 0 -1.1665268 0.044190027 0.06520427375340572 0 +138 0 -1.4441305 0.027544165 0.04029536470297275 0 +141 0 -1.0223967 0.05627053 0.08355473761206976 0 +144 0 -1.2838186 0.03622594 0.053233125000715482 0 +145 0 ? ? ? 0 +147 0 -1.2453936 0.038669728 0.05689592997689058 0 +150 0 -1.0310725 0.055462856 0.082320562822649976 0 +151 1 0.82610506 0.60896397 0.71557123141561685 1 +152 1 3.3665133 0.9928026 0.010421172246529968 1 +154 0 -0.86977047 0.07240758 0.10843706553410887 0 +156 0 -0.64260596 0.104393534 0.15906315143361818 0 +161 0 -1.5917665 0.021360641 0.031150789028338011 0 +164 0 ? ? ? 0 +167 1 -0.6077202 0.11029223 3.1805968909661946 0 +169 0 -0.5327836 0.123955354 0.19092369905046902 0 +171 0 -1.4220948 0.028605305 0.041870486871704173 0 +173 1 3.9830875 0.99756426 0.0035183221210102803 1 +174 1 2.4820702 0.9666171 0.048983566777766128 1 +176 0 -1.5638075 0.022416964 0.032708844858653462 0 +177 1 1.9551601 0.9195135 0.12105730191804394 1 +179 1 1.2502828 0.76703393 0.38263768901529166 1 +180 0 -1.0310725 0.055462856 0.082320562822649976 0 +181 0 -0.86977047 0.07240758 0.10843706553410887 0 +183 1 4.6097193 0.9991928 0.0011650508130881176 1 +187 1 1.6879743 0.87698346 0.18937845514815768 1 +188 1 4.0834093 0.9979587 0.0029479547476608187 1 +189 0 -0.7842487 0.083223455 0.12535795998339563 0 +191 1 2.9199162 0.9843045 0.022823422665535206 1 +192 0 -1.6744659 0.018513307 0.026959387091493849 0 +196 0 1.8771818 0.90872145 3.4535802682537224 1 +198 0 -0.86977047 0.07240758 0.10843706553410887 0 +199 0 -1.4213306 0.02864281 0.041926191402959292 0 +201 1 0.9668718 0.6662815 0.58579621249125136 1 +202 0 -1.4220948 0.028605305 0.041870486871704173 0 +204 0 -1.4220948 0.028605305 0.041870486871704173 0 +205 1 5.0202584 0.9996087 0.00056464353243897288 1 +206 1 1.9891793 0.92384684 0.11427439935960727 1 +207 0 -1.0310725 0.055462856 0.082320562822649976 0 +209 0 -1.5613161 0.022513531 0.032851363266203151 0 +210 1 4.696958 0.9993079 0.0009988769867311344 1 +211 1 4.180914 0.9982809 0.0024822960988749062 1 +212 0 -1.4220948 0.028605305 0.041870486871704173 0 +216 0 -1.1371732 0.046430733 0.068590355669037276 0 +218 1 3.069959 0.9879121 0.017545384997688002 1 +219 0 -1.7835977 0.015319395 0.02227225268559269 0 +223 1 2.241494 0.949843 0.074239041517820473 1 +226 1 3.1111295 0.9887498 0.016322594041481807 1 +228 0 -1.0310725 0.055462856 0.082320562822649976 0 +233 1 1.3529873 0.7978504 0.32580988742795242 1 +237 1 1.5945845 0.858068 0.2208361301333531 1 +239 1 1.7325007 0.8852139 0.17590197097112495 1 +240 0 -0.530623 0.12437006 0.19160681155537265 0 +241 0 -1.3558741 0.03203831 0.046978143292847489 0 +242 0 -1.5638075 0.022416964 0.032708844858653462 0 +244 0 -1.4220948 0.028605305 0.041870486871704173 0 +246 1 4.803934 0.9994269 0.00082704360545659173 1 +247 1 2.2854147 0.95340997 0.068831383140945432 1 +248 0 -0.58386874 0.1144915 0.17542193934316072 0 +249 0 ? ? ? 0 +250 0 -0.40921193 0.14964475 0.23386241348942843 0 +252 0 2.6653335 0.9756179 5.3580329482658149 1 +254 1 1.8133054 0.89892894 0.15372101867397894 1 +257 0 -1.4213306 0.02864281 0.041926191402959292 0 +258 0 -1.6956862 0.017844876 0.025977189944962023 0 +259 0 2.8757467 0.9830535 5.8828693102806131 1 +260 1 2.377999 0.960154 0.058662279876109329 1 +262 1 3.9742966 0.9975263 0.0035732333875340043 1 +267 1 1.9697149 0.92139435 0.11810934644476595 1 +268 1 1.2987179 0.7819603 0.35483271511531139 1 +269 0 -1.4220948 0.028605305 0.041870486871704173 0 +271 0 -1.9356104 0.011756743 0.017061889077286219 0 +272 1 1.9697149 0.92139435 0.11810934644476595 1 +275 0 ? ? ? 0 +276 0 -1.4213306 0.02864281 0.041926191402959292 0 +277 0 -1.1371732 0.046430733 0.068590355669037276 0 +278 0 -1.4220948 0.028605305 0.041870486871704173 0 +279 1 1.0373065 0.69332975 0.52838642550680448 1 +280 0 -1.6956862 0.017844876 0.025977189944962023 0 +283 1 2.2490048 0.9504708 0.073285781412789436 1 +284 1 4.1705155 0.9982491 0.0025282091345333363 1 +285 1 3.5483851 0.9947685 0.0075672700225211196 1 +288 1 0.9313418 0.65219545 0.61662371057560283 1 +290 0 -0.86977047 0.07240758 0.10843706553410887 0 +291 0 -1.4220948 0.028605305 0.041870486871704173 0 +293 1 2.6767123 0.9760911 0.034912311406591241 1 +296 0 1.4451346 0.8228175 2.4966920122534031 1 +297 0 ? ? ? 0 +299 1 1.3851664 0.80685604 0.30961681222475879 1 +300 1 1.9310564 0.9163082 0.12609512677322965 1 +301 0 -1.4220948 0.028605305 0.041870486871704173 0 +303 0 -1.4220948 0.028605305 0.041870486871704173 0 +304 1 2.5191388 0.96866506 0.045930185624366078 1 +308 1 1.9636075 0.9206101 0.11933787110747912 1 +309 0 -0.79236186 0.0821374 0.123649887716065 0 +311 0 -0.86977047 0.07240758 0.10843706553410887 0 +312 1 -0.9134297 0.06739943 3.8911198571593064 0 +314 0 -0.75863177 0.086738825 0.1309005925984742 0 +316 1 2.5552619 0.9705435 0.043135212959228432 1 +317 1 3.6669497 0.99575216 0.0061413968925367029 1 +319 0 -0.1082225 0.23038474 0.37779068446081138 0 +321 0 ? ? ? 0 +323 1 3.2223158 0.9907359 0.013427581534140124 1 +327 0 -1.1371732 0.046430733 0.068590355669037276 0 +328 1 2.7085183 0.97736675 0.03302807661224369 1 +329 1 1.0686762 0.70497465 0.50435671120337722 1 +331 0 -2.0988872 0.008839138 0.012808873813253645 0 +332 0 -0.84503835 0.07539477 0.11309057484193338 0 +333 1 2.8448353 0.9821202 0.026028467234020466 1 +336 1 3.314216 0.9921121 0.011424952948283886 1 +338 0 -0.75863177 0.086738825 0.1309005925984742 0 +343 0 -0.86977047 0.07240758 0.10843706553410887 0 +344 1 1.1524243 0.7347616 0.44465187277066204 1 +346 0 -0.7391351 0.08950382 0.13527512617948156 0 +347 0 -0.14602692 0.21876715 0.35617548289792383 0 +348 1 -0.24912481 0.18925868 2.4015686285382234 0 +349 1 1.8527154 0.9050757 0.14388958542305713 1 +350 0 -1.430436 0.028199043 0.041267241894747161 0 +352 0 0.33406886 0.3952037 0.7254788042723842 1 +353 1 4.66979 0.9992739 0.0010479268242788676 1 +354 0 -1.1371732 0.046430733 0.068590355669037276 0 +355 0 -1.7033824 0.017608354 0.025629802232366293 0 +358 1 2.1114182 0.9377049 0.09279409146249365 1 +360 1 3.333546 0.99237466 0.011043200045097991 1 +361 1 3.5392418 0.99468386 0.0076900250927252559 1 +366 1 4.2243857 0.99840766 0.0022990892087274254 1 +368 0 -0.7202434 0.09225857 0.1396466944003405 0 +370 0 -0.89812446 0.069117405 0.10332887081999793 0 +371 0 -0.7202434 0.09225857 0.1396466944003405 0 +373 0 -1.6798961 0.018339954 0.026704597149590482 0 +376 0 -1.1371732 0.046430733 0.068590355669037276 0 +377 0 -0.75863177 0.086738825 0.1309005925984742 0 +378 0 -1.6366836 0.019764751 0.028800068563565269 0 +379 0 -1.3054267 0.034917682 0.051276091247660721 0 +381 1 2.8214107 0.9813795 0.027116946380001044 1 +383 0 -1.0223967 0.05627053 0.08355473761206976 0 +384 0 -1.0223967 0.05627053 0.08355473761206976 0 +387 0 -1.0608913 0.052769355 0.078212339524604005 0 +388 0 -1.0875129 0.050469454 0.074713683444958351 0 +389 0 -1.4564972 0.0269655 0.039437137813857957 0 +391 1 4.2245107 0.998408 0.0022985724379935217 1 +392 0 -1.4213306 0.02864281 0.041926191402959292 0 +395 0 -1.4213306 0.02864281 0.041926191402959292 0 +396 0 -1.6956862 0.017844876 0.025977189944962023 0 +398 0 -0.76604486 0.08570795 0.1292730157853999 0 +399 0 -0.31279233 0.17261474 0.27336883604144874 0 +404 0 -0.29005778 0.1784211 0.28352895566338898 0 +406 0 -0.99532974 0.05886194 0.087521722324747322 0 +409 0 -1.3234932 0.03385891 0.04969420739241856 0 +413 0 -1.8128492 0.014559762 0.021159712617692564 0 +414 1 3.088689 0.98830056 0.016978234021855474 1 +415 0 0.89851934 0.63894117 1.4696941647367558 1 +416 1 1.3135263 0.78638375 0.34669458862526176 1 +418 0 -0.8207073 0.07844358 0.11785559993419296 0 +419 0 -1.2815632 0.036365185 0.053441577382749392 0 +422 0 -0.16338588 0.21357587 0.34662050567616409 0 +423 0 -1.4396573 0.027756434 0.040610312397176378 0 +428 0 -1.1371732 0.046430733 0.068590355669037276 0 +429 0 -1.2838186 0.03622594 0.053233125000715482 0 +430 0 -0.099619985 0.23308791 0.38286688582511008 0 +434 0 3.0534105 0.9875583 6.3286731752557559 1 +436 1 1.6714998 0.8738119 0.1946053406246547 1 +439 0 -1.4341952 0.028017787 0.040998182116625835 0 +440 1 1.0110853 0.6834023 0.54919299195292437 1 +441 0 0.24502417 0.35832182 0.64007815947259183 1 +442 0 -0.2995455 0.17597957 0.27924798740938434 0 +449 1 4.4427314 0.9989164 0.0015641699264788977 1 +450 0 -1.1997567 0.041777935 0.061568059882899671 0 +451 0 -1.4341952 0.028017787 0.040998182116625835 0 +452 0 -1.1939292 0.042191643 0.062191071963889309 0 +453 1 2.9330091 0.9846576 0.022305979872484345 1 +454 0 -0.57314914 0.11642373 0.17857342756288586 0 +455 1 0.024765657 0.2745987 1.8646033630045489 1 +456 1 2.6778674 0.9761386 0.034842099292583813 1 +457 1 1.9492228 0.91873455 0.12228001033829841 1 +464 0 -1.3033534 0.03504121 0.051460762596864121 0 +465 1 2.2571466 0.9511429 0.072265975739678884 1 +466 1 2.733724 0.97833025 0.031606537500566641 1 +467 1 3.3921504 0.9931188 0.0099617542902060343 1 +474 0 -1.4341952 0.028017787 0.040998182116625835 0 +480 0 -1.0890517 0.05033945 0.074516170061152395 0 +482 1 1.292429 0.78006184 0.35833959375426 1 +483 1 3.9113114 0.9972362 0.0039928520518782568 1 +484 0 -1.3286604 0.033561837 0.04925066941782065 0 +487 1 2.6215835 0.97371197 0.038433021306377303 1 +489 1 -0.40447402 0.15071201 2.7301336762416821 0 +492 0 -1.1864744 0.04272658 0.062997043465874758 0 +493 1 4.5688677 0.99913245 0.0012521469675214095 1 +495 0 -1.0515323 0.053601168 0.079479801859957872 0 +497 0 -0.8915211 0.06987107 0.10449738225465412 0 +501 0 -1.5510771 0.022914685 0.033443557557539826 0 +502 0 -1.5642427 0.022400135 0.03268400921459963 0 +504 0 -0.86977047 0.07240758 0.10843706553410887 0 +507 0 0.02898837 0.2760858 0.46610936701559874 1 +510 0 -0.86977047 0.07240758 0.10843706553410887 0 +513 0 -1.0515323 0.053601168 0.079479801859957872 0 +514 1 4.6778398 0.99928415 0.0010331256449830988 1 +517 0 -0.75863177 0.086738825 0.1309005925984742 0 +519 1 2.9560733 0.98526055 0.02142280786530798 1 +520 0 -1.2458643 0.03863886 0.056849606622682597 0 +521 0 -1.6647452 0.01882765 0.027421517416780897 0 +522 1 0.36196297 0.40703 1.2967930114269521 1 +523 1 2.9920366 0.98615456 0.020114322425649339 1 +527 0 -1.8191407 0.0144012915 0.02092772931478808 0 +528 0 -1.3119496 0.034531783 0.050699328598186447 0 +529 0 -1.1864744 0.04272658 0.062997043465874758 0 +531 0 -0.99532974 0.05886194 0.087521722324747322 0 +532 0 -1.0310725 0.055462856 0.082320562822649976 0 +533 0 -1.4213306 0.02864281 0.041926191402959292 0 +534 0 -1.2838186 0.03622594 0.053233125000715482 0 +535 0 -1.2725202 0.036928657 0.054285420658928121 0 +538 0 -1.5510771 0.022914685 0.033443557557539826 0 +539 0 -1.7928525 0.015074929 0.021914120703602172 0 +540 0 -1.3407189 0.032878313 0.048230668599692793 0 +541 0 -1.1665268 0.044190027 0.06520427375340572 0 +544 0 -1.1073365 0.04881883 0.072207940353225586 0 +546 1 5.8427505 0.9999083 0.00013226072090480578 1 +547 0 -0.6606966 0.10144569 0.15432238949158586 0 +548 0 -0.8099388 0.07982862 0.12002550968300212 0 +549 1 2.620321 0.97365487 0.038517627543231944 1 +557 0 -1.430436 0.028199043 0.041267241894747161 0 +558 0 -1.2838186 0.03622594 0.053233125000715482 0 +559 0 -1.6744659 0.018513307 0.026959387091493849 0 +560 0 -1.9356104 0.011756743 0.017061889077286219 0 +561 0 -1.9356104 0.011756743 0.017061889077286219 0 +563 0 -1.4213306 0.02864281 0.041926191402959292 0 +565 1 4.720941 0.99933654 0.00095748710537352157 1 +566 0 -1.5908056 0.021396121 0.031203093550825723 0 +569 1 1.6528711 0.8701416 0.20067785985732145 1 +577 0 -1.1371732 0.046430733 0.068590355669037276 0 +578 0 -1.1371732 0.046430733 0.068590355669037276 0 +581 1 3.807909 0.9966847 0.0047909556972169825 1 +582 1 5.2861357 0.99975526 0.00035312359513076151 1 +584 0 -1.8085837 0.014668175 0.021318439722779681 0 +586 1 4.7991767 0.9994221 0.00083401291374161218 1 +590 1 1.0879864 0.7120135 0.49002353383393893 1 +593 0 -1.3286604 0.033561837 0.04925066941782065 0 +594 1 3.2776961 0.99159133 0.012182431323830867 1 +600 0 -1.4213306 0.02864281 0.041926191402959292 0 +602 0 -1.5510771 0.022914685 0.033443557557539826 0 +604 1 0.6846822 0.5481866 0.86726103131566612 1 +606 0 -1.3096906 0.03466496 0.050898345761446716 0 +607 0 -0.86977047 0.07240758 0.10843706553410887 0 +609 0 -1.4341952 0.028017787 0.040998182116625835 0 +612 1 3.876799 0.99706316 0.0042431984385250402 1 +613 0 -0.4418751 0.1424558 0.22171706383017925 0 +614 0 -0.90927494 0.06786183 0.10138427891782666 0 +617 0 ? ? ? 0 +618 0 -1.5510771 0.022914685 0.033443557557539826 0 +619 0 -1.6744659 0.018513307 0.026959387091493849 0 +621 0 -1.4372448 0.027871577 0.040781181378282121 0 +622 0 -1.7518852 0.016187044 0.023544040400376672 0 +624 0 -1.3150097 0.034352165 0.050430950258448425 0 +627 0 0.26375008 0.3659563 0.65734583136021085 1 +629 0 -1.3033534 0.03504121 0.051460762596864121 0 +633 1 1.90309 0.91244423 0.13219170690331489 1 +634 0 -1.1665268 0.044190027 0.06520427375340572 0 +638 0 -1.3033534 0.03504121 0.051460762596864121 0 +639 0 -1.430436 0.028199043 0.041267241894747161 0 +641 0 -1.4213306 0.02864281 0.041926191402959292 0 +642 0 -1.4213306 0.02864281 0.041926191402959292 0 +644 0 -1.0223967 0.05627053 0.08355473761206976 0 +645 0 -1.4213306 0.02864281 0.041926191402959292 0 +649 0 -1.4213306 0.02864281 0.041926191402959292 0 +652 0 -1.4478418 0.027369251 0.04003589355988043 0 +653 0 -1.5510771 0.022914685 0.033443557557539826 0 +654 0 -1.6956862 0.017844876 0.025977189944962023 0 +656 0 -1.6744659 0.018513307 0.026959387091493849 0 +657 0 -0.8313994 0.07709015 0.11573836535855195 0 +660 0 -1.1371732 0.046430733 0.068590355669037276 0 +661 0 -1.8191407 0.0144012915 0.02092772931478808 0 +665 0 -0.86977047 0.07240758 0.10843706553410887 0 +668 1 0.86882627 0.6267649 0.67400372130852226 1 +670 1 4.0643644 0.9978891 0.003048601567599658 1 +678 0 -0.86977047 0.07240758 0.10843706553410887 0 +679 0 -1.0223967 0.05627053 0.08355473761206976 0 +680 1 4.132329 0.9981273 0.0027042944357023404 1 +681 1 5.229309 0.9997294 0.00039045344918639427 1 +682 0 -1.6992966 0.017733535 0.025813648883325074 0 +683 0 -0.86977047 0.07240758 0.10843706553410887 0 +685 0 -0.86977047 0.07240758 0.10843706553410887 0 +688 0 -1.3033534 0.03504121 0.051460762596864121 0 +689 0 -2.0930195 0.008930336 0.012941624022771943 0 +691 1 3.427192 0.99352884 0.0093662442293995889 1 +692 0 -1.1665268 0.044190027 0.06520427375340572 0 +693 0 -1.5863222 0.021562442 0.031448311386273131 0 +694 0 -1.2274677 0.039863233 0.058688168525858342 0 +696 1 2.633228 0.9742329 0.03766137298091906 1 +697 1 2.479914 0.9664941 0.049167194182930501 1 +698 1 2.3374114 0.95732105 0.062925265922745013 1 +0 0 -1.9011232 0.0018669876 0.0026960111836120209 0 +1 0 1.8526449 0.9000787 3.3230641435037409 1 +2 0 -1.1825227 0.009393854 0.01361652309740408 0 +3 0 2.3748295 0.9670003 4.92140352789633 1 +4 0 -1.6125494 0.0035768833 0.0051696029224548735 0 +7 0 -1.3333379 0.006699842 0.0096983538538219696 0 +12 1 0.8277002 0.4707278 1.0870350333488543 1 +13 0 -0.6462506 0.030863313 0.045227937251427944 0 +14 1 2.724975 0.9847629 0.022151673899992144 1 +15 1 -0.04991475 0.10912297 3.1959732894881858 0 +16 0 -1.4029442 0.0057306015 0.0082912901511172331 0 +17 0 -1.7603207 0.0025643276 0.0037042941962035 0 +19 0 -2.042641 0.0013568318 0.0019588236373629915 0 +22 0 -1.2844096 0.0074769794 0.010827530345711376 0 +23 1 ? ? ? 0 +24 0 -2.206249 0.0009379914 0.001353870637993779 0 +26 0 -0.57611775 0.035970967 0.05285149951615023 0 +27 0 -1.521502 0.004390087 0.0063474999686655596 0 +29 0 -0.5713078 0.03634966 0.053418337058630458 0 +30 0 -0.90429854 0.017468026 0.025423737782211402 0 +33 0 -1.4419535 0.0052497634 0.0075937577864979159 0 +34 0 -1.122178 0.010751031 0.015594438766898965 0 +36 1 2.7772706 0.9864377 0.019700188382701125 1 +38 1 2.694368 0.98368996 0.023724411581766112 1 +39 1 1.2487694 0.69719183 0.52037242225080371 1 +42 1 2.743568 0.9853805 0.021247215714842692 1 +43 1 -0.6878143 0.028175328 5.1494237880472564 0 +47 0 -1.0449967 0.012772651 0.018545733667249785 0 +49 1 2.8997104 0.98968095 0.014964591911313102 1 +53 1 1.3250071 0.7322719 0.44954864042903153 1 +55 1 2.4606464 0.97265667 0.039997449070786041 1 +57 1 -3.021201 0.00014894486 12.712934019354766 0 +58 1 1.1868198 0.6668603 0.58454356881890979 1 +59 1 1.0272144 0.5826045 0.77941118204948179 1 +61 0 -0.74134856 0.025046365 0.036594483851001988 0 +62 1 2.6574285 0.9822957 0.025770722973519931 1 +65 1 -0.88284653 0.018319475 5.7704780673015676 0 +67 1 1.9326409 0.91519696 0.12784584250814737 1 +75 0 -1.0083907 0.013858488 0.020133405426854115 0 +78 0 -0.73608565 0.025338324 0.037026576285435514 0 +80 0 -3.5258896 4.7636506E-05 6.8726587598935077E-05 0 +81 0 -1.3374258 0.006638666 0.0096095032837452393 0 +83 0 -3.9197981 1.9566109E-05 2.8228204115851847E-05 0 +84 1 2.236427 0.95542794 0.065781021058000766 1 +85 1 1.7272661 0.8715655 0.19831897049362454 1 +86 1 1.6768847 0.85827816 0.22048281513809179 1 +87 1 2.53509 0.9767908 0.033878503651427271 1 +89 0 -1.5181754 0.0044230553 0.0063952737673853453 0 +94 0 -1.4795611 0.0048242547 0.0069767706997042886 0 +101 1 1.4058223 0.7665155 0.38361314186141054 1 +103 1 -7.082733 1.5434262E-08 25.949288270251916 0 +107 1 2.3645048 0.96624786 0.049534786110153849 1 +110 0 -0.53554875 0.039287645 0.05782355428088002 0 +114 0 0.15040311 0.16148524 0.25409191552414345 1 +116 0 0.24694793 0.19323485 0.30977932282438059 1 +118 0 -0.80315393 0.021853939 0.031878183716823785 0 +119 0 -0.77819777 0.023092084 0.033705515151814379 0 +124 1 1.8809547 0.90568495 0.14291881398801479 1 +126 1 2.5040572 0.9751473 0.03630792432339245 1 +127 0 -1.619939 0.0035178792 0.0050841748464657156 0 +130 0 -1.0798098 0.011818085 0.017151441839200841 0 +134 0 -1.6966928 0.0029595392 0.0042760431863012247 0 +135 0 -0.9933846 0.014329473 0.020822606354180775 0 +136 0 -1.4029442 0.0057306015 0.0082912901511172331 0 +139 0 ? ? ? 0 +140 0 -0.980721 0.014739183 0.021422411024334745 0 +142 1 1.8831666 0.9061109 0.14224048808348191 1 +143 0 0.27486685 0.20325744 0.32781445549375171 1 +146 1 1.2918617 0.71734005 0.47927091010433526 1 +148 0 -6.545549 5.1939228E-08 7.493246865186658E-08 0 +149 1 2.6907346 0.98355776 0.023918315743113396 1 +153 0 -0.27656806 0.068387 0.10219732612446057 0 +155 1 2.369683 0.9666273 0.048968354509111099 1 +157 0 -1.4795611 0.0048242547 0.0069767706997042886 0 +158 0 ? ? ? 0 +159 1 3.487656 0.9972449 0.0039802625782586875 1 +160 1 3.2480998 0.99527615 0.0068312180617485802 1 +162 0 -1.619939 0.0035178792 0.0050841748464657156 0 +163 0 -0.37329066 0.055712383 0.082701743697177435 0 +165 0 -1.109832 0.011051727 0.016033032851414428 0 +166 1 2.7681372 0.98615885 0.020108044138011833 1 +168 0 -1.619939 0.0035178792 0.0050841748464657156 0 +170 0 -0.980721 0.014739183 0.021422411024334745 0 +172 0 -1.0449967 0.012772651 0.018545733667249785 0 +175 1 2.5971909 0.97976714 0.029489182913312366 1 +178 0 -1.7603207 0.0025643276 0.0037042941962035 0 +182 0 -2.042641 0.0013568318 0.0019588236373629915 0 +184 1 2.7320902 0.9850022 0.021801119238804373 1 +185 0 -1.0912483 0.011520099 0.016716463147040731 0 +186 1 2.2658408 0.95817345 0.061641250724424279 1 +190 1 3.7211063 0.9983722 0.0023503365592100144 1 +193 0 -2.206249 0.0009379914 0.001353870637993779 0 +194 0 -1.619939 0.0035178792 0.0050841748464657156 0 +195 0 -1.7603207 0.0025643276 0.0037042941962035 0 +197 0 -1.6196405 0.0035202436 0.0050875979939417511 0 +200 1 3.2129972 0.9948883 0.00739352893932981 1 +203 0 -1.9011232 0.0018669876 0.0026960111836120209 0 +208 0 -0.97380435 0.014967809 0.021757222688864538 0 +213 1 3.9037445 0.99892193 0.0015561640801682172 1 +214 1 4.018726 0.99916834 0.00120033617091467 1 +215 1 2.800974 0.98713565 0.018679746477681083 1 +217 0 -2.206249 0.0009379914 0.001353870637993779 0 +220 0 -1.3165237 0.0069574118 0.010072503501908827 0 +221 1 2.6084096 0.9802635 0.028758531283106223 1 +222 1 0.5980332 0.34614354 1.5305576556729361 1 +224 1 3.269716 0.99550027 0.0065063926501169782 1 +225 0 -1.0449967 0.012772651 0.018545733667249785 0 +227 1 2.9631462 0.99104625 0.012975708619839675 1 +229 1 3.4247894 0.9968258 0.0045866653656384876 1 +230 1 2.3573015 0.96571314 0.050333382148227815 1 +231 1 2.7643018 0.9860401 0.02028175350878585 1 +232 0 0.76816535 0.43740255 0.82982507208779333 1 +234 0 -0.15507916 0.08808098 0.13302237846197743 0 +235 0 ? ? ? 0 +236 1 3.0107687 0.9919521 0.011657607431410739 1 +238 1 3.528531 0.9974873 0.0036296122783753973 1 +243 0 -0.34570777 0.05908257 0.087859971934407699 0 +245 0 -1.2606221 0.007886493 0.011422906606939225 0 +251 1 2.9806225 0.9913899 0.012475576493565475 1 +253 1 2.743568 0.9853805 0.021247215714842692 1 +255 1 2.5979743 0.97980225 0.029437489014821822 1 +256 0 -0.980721 0.014739183 0.021422411024334745 0 +261 1 2.7529123 0.9856815 0.020806584005713134 1 +263 1 2.7939656 0.98693305 0.018975870433225177 1 +264 1 1.3940068 0.76170474 0.39269621592066295 1 +265 0 -0.9827596 0.014672456 0.021324708280167328 0 +266 1 3.3755076 0.99645334 0.0051258370079349542 1 +270 1 2.9503555 0.9907862 0.01335432806975332 1 +273 1 0.7553908 0.4303146 1.216536303485716 1 +274 0 -1.5004666 0.0046027494 0.0066556927052846817 0 +281 0 -1.4419535 0.0052497634 0.0075937577864979159 0 +282 1 2.123755 0.9432422 0.084299841778636003 1 +286 1 3.587346 0.9977992 0.0031785566488518415 1 +287 0 -1.6966928 0.0029595392 0.0042760431863012247 0 +289 1 2.8430157 0.9882875 0.016997289181149244 1 +292 1 ? ? ? 0 +294 0 ? ? ? 0 +295 1 2.4589546 0.9725548 0.040148547475821666 1 +298 0 -2.438968 0.00055469514 0.00080047796368677921 0 +302 1 3.5234554 0.9974584 0.0036714237349317726 1 +305 1 3.4062707 0.9966907 0.0047822417098379163 1 +306 0 -2.206249 0.0009379914 0.001353870637993779 0 +307 0 -2.206249 0.0009379914 0.001353870637993779 0 +310 0 -1.5532147 0.004087831 0.0059095802868918889 0 +313 0 -0.7803366 0.02298334 0.033544932585152176 0 +315 0 ? ? ? 0 +318 0 -3.1293073 0.000116675794 0.00016833740979690654 0 +320 1 2.220878 0.95390797 0.068078013682622174 1 +322 0 -1.619939 0.0035178792 0.0050841748464657156 0 +324 0 -2.206249 0.0009379914 0.001353870637993779 0 +325 0 -0.46235713 0.046026126 0.067978337794852739 0 +326 1 2.0489488 0.9334879 0.099296784859202333 1 +330 1 2.4539964 0.9722543 0.040594419794004334 1 +334 1 2.9283304 0.99032074 0.014032238242224359 1 +335 0 -0.7803366 0.02298334 0.033544932585152176 0 +337 0 -2.206249 0.0009379914 0.001353870637993779 0 +339 1 2.8106062 0.98740906 0.018280219223136229 1 +340 1 2.6164298 0.98061097 0.028247198764588193 1 +341 0 -2.206249 0.0009379914 0.001353870637993779 0 +342 0 -0.8815892 0.018370625 0.026749672430821506 0 +345 0 -0.7803366 0.02298334 0.033544932585152176 0 +351 0 -1.4795611 0.0048242547 0.0069767706997042886 0 +356 1 0.31865388 0.21974646 2.186088199604681 1 +357 1 2.8797574 0.9892103 0.015650822159910229 1 +359 1 2.3431916 0.9646419 0.051934575919976009 1 +362 0 -0.5627002 0.037036937 0.054447633744274623 0 +363 0 0.90663147 0.51526546 1.0447332230044488 1 +364 0 -1.4795611 0.0048242547 0.0069767706997042886 0 +365 0 -1.1653385 0.00976205 0.014152854000961439 0 +367 1 3.28966 0.9956976 0.0062204167742769927 1 +369 0 -0.4955235 0.04284612 0.063177213251293107 0 +372 0 -1.2144351 0.008746205 0.012673610575464275 0 +374 0 -1.122178 0.010751031 0.015594438766898965 0 +375 0 -0.7803366 0.02298334 0.033544932585152176 0 +380 0 -0.7803366 0.02298334 0.033544932585152176 0 +382 0 -0.78532875 0.02273147 0.033173060073456939 0 +385 0 -0.40422887 0.052147783 0.077265954445667695 0 +386 1 2.6374905 0.98149514 0.026946967954278624 1 +390 0 -1.0248165 0.013360429 0.019404943898702308 0 +393 0 -0.47379893 0.044904478 0.066283066056692333 0 +394 0 -0.50511205 0.04196657 0.061852094082455338 0 +397 0 -1.0786035 0.01184995 0.017197964153833081 0 +400 1 2.466826 0.9730255 0.039450479645981737 1 +401 0 -0.980721 0.014739183 0.021422411024334745 0 +402 0 -0.20391515 0.07961323 0.11968785079951311 0 +403 0 0.062409937 0.13634409 0.21147145375176429 1 +405 0 -1.0449967 0.012772651 0.018545733667249785 0 +407 0 -1.0449967 0.012772651 0.018545733667249785 0 +408 0 -0.13958576 0.09093303 0.13754151844249329 0 +410 0 -1.0449967 0.012772651 0.018545733667249785 0 +411 0 ? ? ? 0 +412 1 2.659246 0.9823669 0.025666115054241719 1 +417 0 -1.0449967 0.012772651 0.018545733667249785 0 +420 0 -0.0057860482 0.11919814 0.18311058512882616 0 +421 1 3.119992 0.99370074 0.0091166515452388245 1 +424 0 -0.980721 0.014739183 0.021422411024334745 0 +425 1 3.5934417 0.99782926 0.003135122083211179 1 +426 0 0.5049288 0.30019814 0.51498159115469289 1 +427 1 2.286945 0.9600429 0.058829229246299852 1 +431 0 -2.5609357 0.00042116738 0.00060774408107747006 0 +432 0 -1.3668607 0.006214253 0.0089932443658636233 0 +433 0 -0.35668075 0.057719547 0.085771579105318782 0 +435 1 3.2349212 0.9951341 0.0070371226756000792 1 +437 0 -1.0786035 0.01184995 0.017197964153833081 0 +438 0 -0.5534828 0.037786756 0.055571437497170646 0 +443 0 -0.5114841 0.041391645 0.060986580130918226 0 +444 0 -1.480155 0.004817818 0.0069674393214521364 0 +445 0 -0.8815892 0.018370625 0.026749672430821506 0 +446 0 -0.7803366 0.02298334 0.033544932585152176 0 +447 0 -1.1759539 0.009532947 0.013819109069520788 0 +448 0 -0.47379893 0.044904478 0.066283066056692333 0 +458 0 -0.9770439 0.014860296 0.021599764813200976 0 +459 0 -0.76924855 0.02355256 0.034385704749194265 0 +460 0 -1.0172365 0.0135880355 0.019737796109547713 0 +461 0 0.35157394 0.23276228 0.38225443961881339 1 +462 0 -1.0877125 0.011611408 0.016849735187545328 0 +463 0 -0.9526668 0.015688363 0.022812944660740908 0 +468 0 -1.0786035 0.01184995 0.017197964153833081 0 +469 0 -0.93669677 0.0162553 0.023644137598057682 0 +470 0 -0.90429854 0.017468026 0.025423737782211402 0 +471 0 -1.0877125 0.011611408 0.016849735187545328 0 +472 0 -0.85943127 0.019295359 0.028109389806990844 0 +473 0 -1.0786035 0.01184995 0.017197964153833081 0 +475 0 -0.980721 0.014739183 0.021422411024334745 0 +476 0 -0.90481013 0.017448204 0.025394632162802466 0 +477 0 -1.0786035 0.01184995 0.017197964153833081 0 +478 0 -0.7635906 0.023848295 0.034822718884220796 0 +479 1 2.96926 0.991168 0.012798452022955088 1 +481 0 0.07424037 0.13952173 0.21678934001995845 1 +485 0 0.08666081 0.14292437 0.22250557602623472 1 +486 0 -0.90429854 0.017468026 0.025423737782211402 0 +488 1 1.6308237 0.845144 0.24273096339362343 1 +490 0 -0.7803366 0.02298334 0.033544932585152176 0 +491 1 2.796015 0.98699266 0.018888743217612237 1 +494 0 0.68416816 0.39139366 0.71641873560549718 1 +496 0 -0.47379893 0.044904478 0.066283066056692333 0 +498 0 -1.4029442 0.0057306015 0.0082912901511172331 0 +499 0 -1.4029442 0.0057306015 0.0082912901511172331 0 +500 0 -2.042641 0.0013568318 0.0019588236373629915 0 +503 0 -1.7603207 0.0025643276 0.0037042941962035 0 +505 0 -0.33610684 0.060299855 0.089727623138152302 0 +506 1 2.9779997 0.99133915 0.012549392549419343 1 +508 0 -1.1759539 0.009532947 0.013819109069520788 0 +509 0 -0.8815892 0.018370625 0.026749672430821506 0 +511 0 -1.521502 0.004390087 0.0063474999686655596 0 +512 0 -1.1759539 0.009532947 0.013819109069520788 0 +515 1 3.2177675 0.9949428 0.0073145312074519853 1 +516 0 -0.47379893 0.044904478 0.066283066056692333 0 +518 0 -0.7724405 0.023387302 0.034141558324792751 0 +524 0 -1.2844096 0.0074769794 0.010827530345711376 0 +525 0 -0.933973 0.016353983 0.023788866743238088 0 +526 0 -1.0786035 0.01184995 0.017197964153833081 0 +530 1 2.645506 0.9818212 0.026467806701134522 1 +536 0 -1.9011232 0.0018669876 0.0026960111836120209 0 +537 0 -1.6744592 0.003111504 0.0044959496325240305 0 +542 0 -0.37155762 0.055918697 0.083016987421194607 0 +543 0 -1.4029442 0.0057306015 0.0082912901511172331 0 +545 0 -1.521502 0.004390087 0.0063474999686655596 0 +550 0 -1.2844096 0.0074769794 0.010827530345711376 0 +551 0 -2.206249 0.0009379914 0.001353870637993779 0 +552 0 -1.0870707 0.011628059 0.016874040086919551 0 +553 0 1.0316805 0.58505577 1.2690106435607837 1 +554 0 -0.980721 0.014739183 0.021422411024334745 0 +555 0 0.06918905 0.13815743 0.21450372945092822 1 +556 0 -0.38651156 0.054161884 0.080334813089593907 0 +562 0 -2.206249 0.0009379914 0.001353870637993779 0 +564 0 -1.6200049 0.0035173572 0.0050834190921056925 0 +567 0 -0.51049584 0.041480314 0.061120032715017494 0 +568 1 2.2669675 0.9582753 0.061487884584456066 1 +570 1 2.5896778 0.979428 0.029988664192482247 1 +571 1 3.3817437 0.9965028 0.0050542119501423713 1 +572 0 -1.2844096 0.0074769794 0.010827530345711376 0 +573 0 -1.0449967 0.012772651 0.018545733667249785 0 +574 1 2.4260175 0.9704975 0.043203614712199016 1 +575 0 -1.6744592 0.003111504 0.0044959496325240305 0 +576 0 -1.521502 0.004390087 0.0063474999686655596 0 +579 0 -2.206249 0.0009379914 0.001353870637993779 0 +580 0 -1.3069996 0.007107645 0.010290778994843274 0 +583 0 -0.980721 0.014739183 0.021422411024334745 0 +585 0 -0.7803366 0.02298334 0.033544932585152176 0 +587 0 -1.3668607 0.006214253 0.0089932443658636233 0 +588 1 2.14996 0.9463294 0.079585625448312408 1 +589 0 -1.1759539 0.009532947 0.013819109069520788 0 +591 1 2.3894417 0.96803755 0.046865090861326217 1 +592 1 2.4784293 0.973705 0.038443353952416652 1 +595 0 -1.521502 0.004390087 0.0063474999686655596 0 +596 0 -1.2144351 0.008746205 0.012673610575464275 0 +597 0 -1.4302268 0.00538993 0.0077970573162372174 0 +598 0 -1.2844096 0.0074769794 0.010827530345711376 0 +599 0 0.12021219 0.15246218 0.23865035481609512 1 +601 0 -0.70943594 0.026868323 0.039293061620344245 0 +603 1 1.283247 0.71337754 0.48726230963954414 1 +605 1 2.9619548 0.99102235 0.013010503098224284 1 +608 1 2.680125 0.9831656 0.024493624851198367 1 +610 1 2.6469262 0.98187834 0.026383816579135237 1 +611 1 2.3583596 0.9657922 0.050215314135409014 1 +615 0 -1.0116202 0.013759139 0.019988068422653903 0 +616 0 -1.2844096 0.0074769794 0.010827530345711376 0 +620 0 -1.2844096 0.0074769794 0.010827530345711376 0 +623 0 -0.7803366 0.02298334 0.033544932585152176 0 +625 0 0.12630653 0.15424964 0.24169620590652488 1 +626 1 1.7800001 0.88432056 0.1773586683318118 1 +628 0 -0.8815892 0.018370625 0.026749672430821506 0 +630 0 -0.3736648 0.055667937 0.082633839441836929 0 +631 0 -1.521502 0.004390087 0.0063474999686655596 0 +632 0 -0.7803366 0.02298334 0.033544932585152176 0 +635 0 -0.5025347 0.04220128 0.062205588221011146 0 +636 1 2.4077213 0.9692908 0.044998547307833543 1 +637 0 0.18569934 0.17257519 0.27329987919203769 1 +640 0 -0.9471856 0.015880719 0.023094905413264373 0 +643 0 -0.7803366 0.02298334 0.033544932585152176 0 +646 0 -0.42931885 0.049416363 0.073114527422748377 0 +647 0 -0.73560995 0.025364876 0.037065879123510143 0 +648 1 1.9894372 0.92463875 0.11303827306643183 1 +650 0 -0.1209569 0.09447212 0.14316903073583315 0 +651 0 -0.68340445 0.02844938 0.041638930565002674 0 +655 0 -1.2844096 0.0074769794 0.010827530345711376 0 +658 1 3.0674646 0.9929127 0.010261203719457398 1 +659 0 -0.7803366 0.02298334 0.033544932585152176 0 +662 0 -0.80848724 0.02159788 0.03150056627259288 0 +663 0 -0.80848724 0.02159788 0.03150056627259288 0 +664 0 -1.1712708 0.009633353 0.013965365744220474 0 +666 0 -0.5658891 0.03678087 0.05406405205376117 0 +667 0 -1.619939 0.0035178792 0.0050841748464657156 0 +669 1 2.9674554 0.99113226 0.012850507501880569 1 +671 0 -1.150161 0.010099112 0.014644009725377232 0 +672 0 -1.4795611 0.0048242547 0.0069767706997042886 0 +673 0 -0.49901578 0.042523757 0.062691403137252244 0 +674 0 -1.0449967 0.012772651 0.018545733667249785 0 +675 0 -0.5305139 0.039719187 0.058471742511275952 0 +676 0 -0.93669677 0.0162553 0.023644137598057682 0 +677 0 -1.1759539 0.009532947 0.013819109069520788 0 +684 0 -0.7803366 0.02298334 0.033544932585152176 0 +686 0 -0.7803366 0.02298334 0.033544932585152176 0 +687 0 -1.0420065 0.012858108 0.018670621528273802 0 +690 0 -0.73560995 0.025364876 0.037065879123510143 0 +695 0 -0.8815892 0.018370625 0.026749672430821506 0 diff --git a/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer-out.txt b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer-out.txt new file mode 100644 index 0000000000..614a33f0a0 --- /dev/null +++ b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer-out.txt @@ -0,0 +1,38 @@ +maml.exe TrainTest test=%Data% tr=LdSvm{iter=1000 bias=-} dout=%Output% data=%Data% out=%Output% seed=1 +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Warning: Skipped 16 rows with missing feature/label values +Training calibrator. +Warning: The predictor produced non-finite prediction values on 16 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3499 (239.0/(239.0+444.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 232 | 7 | 0.9707 + negative || 19 | 425 | 0.9572 + ||====================== +Precision || 0.9243 | 0.9838 | +OVERALL 0/1 ACCURACY: 0.961933 +LOG LOSS/instance: 0.170625 +Test-set entropy (prior Log-Loss/instance): 0.934003 +LOG-LOSS REDUCTION (RIG): 0.817319 +AUC: 0.986788 + +OVERALL RESULTS +--------------------------------------- +AUC: 0.986788 (0.0000) +Accuracy: 0.961933 (0.0000) +Positive precision: 0.924303 (0.0000) +Positive recall: 0.970711 (0.0000) +Negative precision: 0.983796 (0.0000) +Negative recall: 0.957207 (0.0000) +Log-loss: 0.170625 (0.0000) +Log-loss reduction: 0.817319 (0.0000) +F1 Score: 0.946939 (0.0000) +AUPRC: 0.978991 (0.0000) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer-rp.txt b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer-rp.txt new file mode 100644 index 0000000000..11096c91be --- /dev/null +++ b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer-rp.txt @@ -0,0 +1,4 @@ +LdSvm +AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC /bias /iter Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.986788 0.961933 0.924303 0.970711 0.983796 0.957207 0.170625 0.817319 0.946939 0.978991 - 1000 LdSvm %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=LdSvm{iter=1000 bias=-} dout=%Output% data=%Data% out=%Output% seed=1 /bias:-;/iter:1000 + diff --git a/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer.txt b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer.txt new file mode 100644 index 0000000000..c9fa789b7c --- /dev/null +++ b/test/BaselineOutput/Common/LdSvm/net9.0/LDSVM-nob-TrainTest-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +0 0 -1.4465218 0.006447295 0.0093315950043344458 0 +1 0 1.5368496 0.91324556 3.5269185736212378 1 +2 0 -1.130633 0.013994661 0.020332636623517589 0 +3 0 2.322816 0.9866292 6.2247691143847677 1 +4 0 -1.4694004 0.0060941684 0.0088189259247697643 0 +5 1 3.3938828 0.9990469 0.0013756569522042779 1 +6 0 0.9102346 0.6902808 1.6909672512393978 1 +7 0 -1.225957 0.011083427 0.016079277270465417 0 +8 0 -1.316685 0.0088719865 0.012856687859128777 0 +9 0 -0.94465667 0.022005446 0.03210166376635272 0 +10 0 -0.9263915 0.023000715 0.033570588948334805 0 +11 0 -1.1461768 0.0134730805 0.019569676334993513 0 +12 1 0.5397904 0.47093934 1.086386856925831 1 +13 0 -0.81922036 0.029787239 0.043626939306907829 0 +14 1 2.3675165 0.98801416 0.017396374226630464 1 +15 1 -0.095254965 0.15580857 2.6821535246945882 0 +16 0 -1.2618061 0.01015102 0.014719663419588126 0 +17 0 -1.4161462 0.0069477297 0.010058437491048256 0 +18 1 2.5378735 0.99210846 0.011430240133534886 1 +19 0 -1.4653347 0.0061554867 0.0089079346640132586 0 +20 1 1.698243 0.94012696 0.08907250222894815 1 +21 1 2.0263994 0.9725304 0.040184710880495383 1 +22 0 -1.2114133 0.011485413 0.016665839255833863 0 +23 1 ? ? ? 0 +24 0 -1.2416301 0.010665795 0.015470137413992897 0 +25 1 1.5185649 0.9095885 0.13671405565591807 1 +26 0 -0.63984776 0.045693673 0.067475657558432545 0 +27 0 -1.2988322 0.009269498 0.01343542569044209 0 +28 0 -1.1461768 0.0134730805 0.019569676334993513 0 +29 0 -0.77650315 0.033002906 0.048416540061604084 0 +30 0 -0.9095942 0.023954712 0.034980005022197523 0 +31 0 -1.3151383 0.008905744 0.012905826640471321 0 +32 1 1.3522481 0.8695014 0.20173972467480639 1 +33 0 -1.3616288 0.007944517 0.011507285721941056 0 +34 0 -1.117744 0.014442215 0.020987632993847882 0 +35 0 -1.1461768 0.0134730805 0.019569676334993513 0 +36 1 2.2732155 0.9849072 0.021940283172727205 1 +37 0 -1.2952731 0.009350831 0.013553866242308086 0 +38 1 2.6219492 0.99358284 0.009287830780025973 1 +39 1 1.2902459 0.85106033 0.23266668805413909 1 +40 0 ? ? ? 0 +41 1 0.6595702 0.54497576 0.87573603962120883 1 +42 1 2.791701 0.9957768 0.0061057314127129642 1 +43 1 -0.7793541 0.032778226 4.9311184208268601 0 +44 1 2.0215683 0.97220886 0.040661816690197673 1 +45 0 -1.1576908 0.013099131 0.019022917233253389 0 +46 1 2.9830015 0.9973667 0.0038040214819385918 1 +47 0 -1.0646522 0.016439164 0.023913804967856618 0 +48 0 -1.4694004 0.0060941684 0.0088189259247697643 0 +49 1 2.8656814 0.99648154 0.0050850189187078036 1 +50 1 1.128352 0.7927918 0.33498608382420991 1 +51 1 0.8557086 0.6606831 0.5979696631659619 1 +52 1 1.7965511 0.9524545 0.070277908934009301 1 +53 1 1.0211413 0.74577665 0.42318446097001949 1 +54 1 1.1301751 0.7935328 0.33363825847496043 1 +55 1 1.8724201 0.96027833 0.058475469955889613 1 +56 1 3.2046094 0.9984776 0.00219806404939647 1 +57 1 -0.9766559 0.020362355 5.6179517775023706 0 +58 1 0.6906531 0.5639973 0.82623976645531594 1 +59 1 0.5961009 0.5057848 0.98340438619611792 1 +60 1 1.4791523 0.9012306 0.15003183800390588 1 +61 0 -0.8840541 0.025479626 0.037235746971575577 0 +62 1 2.346089 0.9873689 0.018338917623867506 1 +63 1 0.67293215 0.5531722 0.85419951541088113 1 +64 0 -1.0646522 0.016439164 0.023913804967856618 0 +65 1 -0.07758474 0.16165413 2.6290177317939487 0 +66 0 -1.4161462 0.0069477297 0.010058437491048256 0 +67 1 1.8333136 0.95641327 0.064293949102047873 1 +68 1 2.2789576 0.98511726 0.021632638839248739 1 +69 0 -1.1185982 0.014412121 0.020943581237724104 0 +70 0 -1.0483546 0.017104952 0.024890718331546519 0 +71 1 1.1304778 0.79365563 0.33341493511960385 1 +72 0 -0.70869404 0.038806155 0.057100685166792849 0 +73 1 2.5126984 0.99160486 0.012162745897967646 1 +74 1 1.1138558 0.78682977 0.34587655204190176 1 +75 0 -0.92996174 0.022802776 0.033278329247217478 0 +76 0 -0.8348704 0.028686864 0.041991622629636863 0 +77 0 -0.63210064 0.046537984 0.068752629505610646 0 +78 0 -0.8731325 0.026160212 0.03824364948616471 0 +79 0 -1.4683211 0.006110387 0.0088424679588101143 0 +80 0 -1.1583824 0.013076996 0.018990560055578542 0 +81 0 -1.292458 0.009415659 0.013648279806877021 0 +82 0 -1.0896415 0.015467499 0.022489262853638857 0 +83 0 -1.7367461 0.0031516228 0.0045540106703456096 0 +84 1 2.4358482 0.9898619 0.014700824279765249 1 +85 1 1.855396 0.9586379 0.06094212671107798 1 +86 1 1.5558441 0.91690224 0.12516016639190308 1 +87 1 2.8115852 0.99597895 0.0058128415032152186 1 +88 0 -1.4161462 0.0069477297 0.010058437491048256 0 +89 0 -1.5202216 0.00537705 0.0077783745497683318 0 +90 0 -1.2416301 0.010665795 0.015470137413992897 0 +91 0 -1.0939798 0.015304665 0.022250671629517431 0 +92 0 -1.4161462 0.0069477297 0.010058437491048256 0 +93 0 -1.0646522 0.016439164 0.023913804967856618 0 +94 0 -1.3151383 0.008905744 0.012905826640471321 0 +95 0 -1.2416301 0.010665795 0.015470137413992897 0 +96 0 -1.0309523 0.017845122 0.025977551104063305 0 +97 0 -1.4465218 0.006447295 0.0093315950043344458 0 +98 1 1.786953 0.95136595 0.071927707099913571 1 +99 1 2.8233905 0.9960944 0.0056456135274451213 1 +100 1 0.8785555 0.67325515 0.57077474389859006 1 +101 1 1.3359516 0.86485124 0.20947609976473319 1 +102 0 -1.2974266 0.009301535 0.013482077873873375 0 +103 1 -1.6196158 0.0042082923 7.8925493537675893 0 +104 1 1.7452879 0.9463609 0.079537647794694813 1 +105 1 0.15215418 0.2541169 1.9764358109174704 1 +106 1 3.4643967 0.99919957 0.0011552399157125351 1 +107 1 2.7960794 0.99582213 0.0060400159731946786 1 +108 0 -0.83246773 0.028853199 0.042238700614129362 0 +109 1 2.423801 0.989558 0.01514385284049204 1 +110 0 -0.35937724 0.08753301 0.13215573015212626 0 +111 1 1.7971286 0.95251924 0.070179863929785383 1 +112 1 2.0087795 0.9713398 0.041951986184448342 1 +113 1 2.1212506 0.97815776 0.031860930979857816 1 +114 0 -0.08898843 0.15786165 0.24787083015030675 0 +115 0 -0.1079554 0.1517144 0.23737802217732704 0 +116 0 0.016546013 0.19579989 0.31437355635618991 1 +117 1 2.4341207 0.9898189 0.014763547257838518 1 +118 0 -1.124218 0.014215678 0.020656058409529759 0 +119 0 -0.7989723 0.03127175 0.045836078193321424 0 +120 0 -1.2139035 0.011415574 0.016563915878113233 0 +121 0 -0.8100147 0.030453498 0.044617998768972537 0 +122 1 2.6066825 0.99333715 0.0096446204297263617 1 +123 1 1.1917268 0.8174033 0.29087999798679592 1 +124 1 1.7512708 0.9471084 0.078398556410385303 1 +125 0 -1.0646522 0.016439164 0.023913804967856618 0 +126 1 2.1918793 0.9815996 0.026793391269283585 1 +127 0 -1.3728285 0.007728782 0.011193587824129689 0 +128 1 2.122015 0.9781982 0.031801328207401759 1 +129 0 -3.9088361 1.4545506E-05 2.0984882019359247E-05 0 +130 0 -1.0483546 0.017104952 0.024890718331546519 0 +131 0 -1.3151383 0.008905744 0.012905826640471321 0 +132 1 3.2652373 0.99868965 0.0018916729445432766 1 +133 0 -1.0670953 0.01634158 0.023770675092729076 0 +134 0 -1.5313979 0.0052309684 0.0075664994594761567 0 +135 0 -0.7877467 0.03212536 0.047107896339556317 0 +136 0 -1.2618061 0.01015102 0.014719663419588126 0 +137 0 -1.0517125 0.016965643 0.024686254694263519 0 +138 0 -1.1537459 0.013226084 0.019208514989069607 0 +139 0 ? ? ? 0 +140 0 -1.0517125 0.016965643 0.024686254694263519 0 +141 0 -0.9785019 0.020271322 0.029545823977617132 0 +142 1 1.27243 0.8453781 0.24233135438110676 1 +143 0 -0.1079554 0.1517144 0.23737802217732704 0 +144 0 -1.1461768 0.0134730805 0.019569676334993513 0 +145 0 ? ? ? 0 +146 1 1.0567243 0.7621264 0.39189783001028217 1 +147 0 -1.1928443 0.012019657 0.017445756275161716 0 +148 0 -2.0253875 0.0015440299 0.0022292857902692985 0 +149 1 2.4897327 0.99111766 0.012871764029188529 1 +150 0 -0.9263915 0.023000715 0.033570588948334805 0 +151 1 1.2475507 0.83714825 0.25644496509411652 1 +152 1 2.9390476 0.99706465 0.0042410423248346301 1 +153 0 -0.64146185 0.04551961 0.06721253592110879 0 +154 0 -0.8886014 0.02520137 0.036823870048860327 0 +155 1 2.2297552 0.98322016 0.024413597764428766 1 +156 0 -0.7366829 0.036300715 0.053345060263684213 0 +157 0 -1.3151383 0.008905744 0.012905826640471321 0 +158 0 ? ? ? 0 +159 1 3.3656669 0.998978 0.0014751611878357862 1 +160 1 2.8545523 0.99638355 0.0052268947939434511 1 +161 0 -1.242877 0.010633245 0.01542267261125157 0 +162 0 -1.3728285 0.007728782 0.011193587824129689 0 +163 0 -0.6702825 0.042515963 0.062679660475101706 0 +164 0 ? ? ? 0 +165 0 -0.97401017 0.020493526 0.029873066958112877 0 +166 1 2.8904924 0.99669063 0.0047823279866827002 1 +167 1 1.248098 0.837333 0.25612657024041963 1 +168 0 -1.3728285 0.007728782 0.011193587824129689 0 +169 0 -0.65122175 0.044480477 0.065642743790444241 0 +170 0 -1.0517125 0.016965643 0.024686254694263519 0 +171 0 -1.2416301 0.010665795 0.015470137413992897 0 +172 0 -1.0646522 0.016439164 0.023913804967856618 0 +173 1 3.5119874 0.99928856 0.001026757742470385 1 +174 1 2.2319386 0.98330915 0.024283027578344871 1 +175 1 2.6796863 0.99443346 0.0080532510881181595 1 +176 0 -1.3151383 0.008905744 0.012905826640471321 0 +177 1 2.0602596 0.97468495 0.036992121356490264 1 +178 0 -1.4161462 0.0069477297 0.010058437491048256 0 +179 1 1.1299858 0.79345596 0.33377794795766014 1 +180 0 -0.9263915 0.023000715 0.033570588948334805 0 +181 0 -0.8886014 0.02520137 0.036823870048860327 0 +182 0 -1.4653347 0.0061554867 0.0089079346640132586 0 +183 1 3.3139293 0.99883837 0.0016768590790372055 1 +184 1 2.392941 0.9887376 0.016340422950879608 1 +185 0 -1.0021135 0.019141525 0.027883105423987049 0 +186 1 2.7014964 0.9947247 0.0076308074344414228 1 +187 1 2.4288478 0.98968637 0.014956685131654837 1 +188 1 2.8906095 0.9966916 0.0047809475577853169 1 +189 0 -0.77104795 0.033436976 0.049064290118216128 0 +190 1 3.4450624 0.99916035 0.0012118686457172167 1 +191 1 2.8183694 0.9960457 0.005716145627458453 1 +192 0 -1.2988322 0.009269498 0.01343542569044209 0 +193 0 -1.2416301 0.010665795 0.015470137413992897 0 +194 0 -1.3728285 0.007728782 0.011193587824129689 0 +195 0 -1.4161462 0.0069477297 0.010058437491048256 0 +196 0 1.8634127 0.9594183 4.6230267744957549 1 +197 0 -1.4985853 0.005671476 0.0082055010281858239 0 +198 0 -0.8886014 0.02520137 0.036823870048860327 0 +199 0 -1.2114133 0.011485413 0.016665839255833863 0 +200 1 2.9217806 0.99693674 0.0044261348585596218 1 +201 1 2.0116594 0.9715378 0.041657924245927638 1 +202 0 -1.2416301 0.010665795 0.015470137413992897 0 +203 0 -1.4465218 0.006447295 0.0093315950043344458 0 +204 0 -1.2416301 0.010665795 0.015470137413992897 0 +205 1 3.769681 0.99962413 0.00054236323274763369 1 +206 1 2.3212264 0.98657715 0.019496216574047962 1 +207 0 -0.9263915 0.023000715 0.033570588948334805 0 +208 0 -0.9263915 0.023000715 0.033570588948334805 0 +209 0 -1.1418804 0.013615301 0.019777674174686889 0 +210 1 3.8264174 0.9996734 0.00047122342085044325 1 +211 1 3.2648933 0.9986885 0.0018933089243565492 1 +212 0 -1.2416301 0.010665795 0.015470137413992897 0 +213 1 3.6728404 0.99952227 0.00068938515685593627 1 +214 1 3.7137108 0.9995683 0.00062296965289850557 1 +215 1 2.529658 0.99194753 0.011664282498920688 1 +216 0 -1.0646522 0.016439164 0.023913804967856618 0 +217 0 -1.2416301 0.010665795 0.015470137413992897 0 +218 1 2.5768757 0.99283016 0.010381156798751557 1 +219 0 -1.1761291 0.0125215165 0.018178781780413589 0 +220 0 -1.2982855 0.009281946 0.013453552581319144 0 +221 1 2.1333241 0.9787878 0.030932005444301781 1 +222 1 0.01927614 0.19686717 2.344705565596565 1 +223 1 2.0440307 0.97367364 0.038489807617213469 1 +224 1 2.8055017 0.99591815 0.0059009094578254748 1 +225 0 -1.0646522 0.016439164 0.023913804967856618 0 +226 1 2.7871819 0.9957294 0.0061743860873786005 1 +227 1 2.7330256 0.99511915 0.0070588121992527569 1 +228 0 -0.9263915 0.023000715 0.033570588948334805 0 +229 1 3.1574626 0.9982893 0.0024701504933073836 1 +230 1 2.0244212 0.97239923 0.040379336847198161 1 +231 1 2.762157 0.9954575 0.0065684148322284388 1 +232 0 0.43547785 0.40737766 0.75481508316172163 1 +233 1 1.7566785 0.94777554 0.077382662054283152 1 +234 0 0.027403058 0.20007019 0.32205467429669943 1 +235 0 ? ? ? 0 +236 1 2.6846595 0.99450123 0.007954935002362596 1 +237 1 2.0117264 0.9715424 0.041651108951427755 1 +238 1 3.552194 0.999356 0.00092943559478244481 1 +239 1 1.8518755 0.95829064 0.061464822742433554 1 +240 0 -0.37127057 0.0852079 0.1284841909589759 0 +241 0 -1.1269433 0.014121364 0.02051803634822633 0 +242 0 -1.3151383 0.008905744 0.012905826640471321 0 +243 0 -0.5598295 0.05516024 0.081858417512951656 0 +244 0 -1.2416301 0.010665795 0.015470137413992897 0 +245 0 -1.17999 0.012403787 0.018006790667367857 0 +246 1 3.7017608 0.9995553 0.00064172398023939646 1 +247 1 1.9196883 0.96451235 0.05212838638294684 1 +248 0 -0.51093537 0.06182471 0.092070590511131534 0 +249 0 ? ? ? 0 +250 0 -0.59787124 0.050449084 0.074682734234965215 0 +251 1 2.3190875 0.98650676 0.019599157715772007 1 +252 0 2.0363448 0.9731811 5.2206076579140177 1 +253 1 2.791701 0.9957768 0.0061057314127129642 1 +254 1 2.346089 0.9873689 0.018338917623867506 1 +255 1 2.6827943 0.9944759 0.0079916838545051432 1 +256 0 -1.0517125 0.016965643 0.024686254694263519 0 +257 0 -1.2114133 0.011485413 0.016665839255833863 0 +258 0 -1.3728285 0.007728782 0.011193587824129689 0 +259 0 2.2774496 0.98506236 6.0649040289959828 1 +260 1 2.5214987 0.9917844 0.01190157086960464 1 +261 1 2.9015005 0.9967794 0.0046538674803215619 1 +262 1 3.2181156 0.9985276 0.0021258091315523894 1 +263 1 2.1179478 0.9779822 0.032119853609565645 1 +264 1 0.999928 0.735684 0.44284192351483948 1 +265 0 -0.8306359 0.028980644 0.042428041342512206 0 +266 1 3.041239 0.9977197 0.0032935266718119955 1 +267 1 1.569662 0.91947377 0.12111967996095568 1 +268 1 2.0937722 0.9766549 0.034079236361848557 1 +269 0 -1.2416301 0.010665795 0.015470137413992897 0 +270 1 2.748885 0.9953064 0.006787414198875746 1 +271 0 -1.4465218 0.006447295 0.0093315950043344458 0 +272 1 1.569662 0.91947377 0.12111967996095568 1 +273 1 0.32582885 0.34378293 1.540430176470527 1 +274 0 -1.2142047 0.011407157 0.016551633475249992 0 +275 0 ? ? ? 0 +276 0 -1.2114133 0.011485413 0.016665839255833863 0 +277 0 -1.0646522 0.016439164 0.023913804967856618 0 +278 0 -1.2416301 0.010665795 0.015470137413992897 0 +279 1 1.8378671 0.95688117 0.063588326479722193 1 +280 0 -1.3728285 0.007728782 0.011193587824129689 0 +281 0 -1.3616288 0.007944517 0.011507285721941056 0 +282 1 2.3653634 0.9879508 0.017488894874168621 1 +283 1 2.1986032 0.9818981 0.026354740846238428 1 +284 1 2.932374 0.99701583 0.0043116782857304265 1 +285 1 3.3162715 0.9988451 0.001667130791306249 1 +286 1 3.3584232 0.9989595 0.0015019320975399703 1 +287 0 -1.5313979 0.0052309684 0.0075664994594761567 0 +288 1 0.93950224 0.705566 0.50314707574137463 1 +289 1 2.967706 0.9972653 0.0039507725975256548 1 +290 0 -0.8886014 0.02520137 0.036823870048860327 0 +291 0 -1.2416301 0.010665795 0.015470137413992897 0 +292 1 ? ? ? 0 +293 1 2.2853177 0.9853465 0.021296958836365919 1 +294 0 ? ? ? 0 +295 1 2.3386807 0.9871379 0.01867643622683381 1 +296 0 1.2135897 0.82534945 2.517456909810909 1 +297 0 ? ? ? 0 +298 0 -1.7178994 0.0033017776 0.0047713395261542143 0 +299 1 1.6048926 0.92570454 0.11137629923606814 1 +300 1 2.0800734 0.9758684 0.035241481755385426 1 +301 0 -1.2416301 0.010665795 0.015470137413992897 0 +302 1 3.283662 0.99874806 0.0018072933433256014 1 +303 0 -1.2416301 0.010665795 0.015470137413992897 0 +304 1 2.099307 0.9769655 0.03362049749270523 1 +305 1 3.1138442 0.99809444 0.0027517653351665072 1 +306 0 -1.2416301 0.010665795 0.015470137413992897 0 +307 0 -1.2416301 0.010665795 0.015470137413992897 0 +308 1 2.0704226 0.9752989 0.036083692614456077 1 +309 0 -0.54896575 0.05657994 0.084027817610603184 0 +310 0 -1.4683211 0.006110387 0.0088424679588101143 0 +311 0 -0.8886014 0.02520137 0.036823870048860327 0 +312 1 0.18371412 0.26922026 1.8931410969312452 1 +313 0 -0.8886014 0.02520137 0.036823870048860327 0 +314 0 -0.7576831 0.03452383 0.050687443773780996 0 +315 0 ? ? ? 0 +316 1 2.0352976 0.9731133 0.039320308845729388 1 +317 1 2.8848646 0.9966443 0.004849366654624771 1 +318 0 -2.2963443 0.000789649 0.0011396727665334838 0 +319 0 0.4430366 0.41190657 0.76588272212659969 1 +320 1 1.9531773 0.9672454 0.048046132405029744 1 +321 0 ? ? ? 0 +322 0 -1.3728285 0.007728782 0.011193587824129689 0 +323 1 2.3755345 0.9882471 0.017056283461848766 1 +324 0 -1.2416301 0.010665795 0.015470137413992897 0 +325 0 -0.4780615 0.0667209 0.099619511386915785 0 +326 1 1.3067058 0.85615593 0.22405451582757585 1 +327 0 -1.0646522 0.016439164 0.023913804967856618 0 +328 1 2.200109 0.9819643 0.026257534066032991 1 +329 1 1.7265749 0.94395816 0.083205174911271079 1 +330 1 1.7580627 0.94794506 0.077124650056060229 1 +331 0 -1.5149018 0.005448002 0.0078812935254501185 0 +332 0 -0.6506068 0.044545274 0.065740582178096446 0 +333 1 2.2095468 0.98237383 0.02565596104920673 1 +334 1 2.6821816 0.99446756 0.0080037895637091391 1 +335 0 -0.8886014 0.02520137 0.036823870048860327 0 +336 1 2.478046 0.9908591 0.013248186580367081 1 +337 0 -1.2416301 0.010665795 0.015470137413992897 0 +338 0 -0.7576831 0.03452383 0.050687443773780996 0 +339 1 2.4499636 0.9902069 0.014198096499880937 1 +340 1 2.015386 0.97179204 0.041280476226848252 1 +341 0 -1.2416301 0.010665795 0.015470137413992897 0 +342 0 -0.9785019 0.020271322 0.029545823977617132 0 +343 0 -0.8886014 0.02520137 0.036823870048860327 0 +344 1 1.8061669 0.9535218 0.068662190167968165 1 +345 0 -0.8886014 0.02520137 0.036823870048860327 0 +346 0 -0.61363167 0.04861104 0.071892812404899059 0 +347 0 -0.4149763 0.07713775 0.11581277846997885 0 +348 1 -0.067022905 0.16523191 2.5974357352865147 0 +349 1 1.4695055 0.8990826 0.15347442889143131 1 +350 0 -1.1986376 0.011850395 0.017198614103561393 0 +351 0 -1.3151383 0.008905744 0.012905826640471321 0 +352 0 0.49842453 0.4455021 0.85074610445850252 1 +353 1 3.334415 0.9988958 0.0015938694218608053 1 +354 0 -1.0646522 0.016439164 0.023913804967856618 0 +355 0 -1.4051663 0.00713796 0.010334828070178312 0 +356 1 0.1424656 0.249594 2.0023448241517841 1 +357 1 2.6753974 0.99437433 0.008139034535833152 1 +358 1 2.1404142 0.97914946 0.03039900024083278 1 +359 1 1.6420252 0.9317892 0.10192445604628746 1 +360 1 3.3180683 0.9988502 0.0016597270056659794 1 +361 1 2.5965574 0.99316907 0.0098887631717976352 1 +362 0 -0.8188577 0.029813215 0.043665566500560457 0 +363 0 0.5482914 0.47619006 0.93288465889381078 1 +364 0 -1.3151383 0.008905744 0.012905826640471321 0 +365 0 -1.1461768 0.0134730805 0.019569676334993513 0 +366 1 3.5748906 0.99939114 0.00087866891024629781 1 +367 1 3.4921196 0.9992527 0.0010785623057035674 1 +368 0 -0.77650315 0.033002906 0.048416540061604084 0 +369 0 -0.63856554 0.045832403 0.067685400859399808 0 +370 0 -0.7563023 0.03463804 0.050858115904229388 0 +371 0 -0.77650315 0.033002906 0.048416540061604084 0 +372 0 -1.1537459 0.013226084 0.019208514989069607 0 +373 0 -1.3214728 0.008768285 0.012705746408436134 0 +374 0 -1.117744 0.014442215 0.020987632993847882 0 +375 0 -0.8886014 0.02520137 0.036823870048860327 0 +376 0 -1.0646522 0.016439164 0.023913804967856618 0 +377 0 -0.7576831 0.03452383 0.050687443773780996 0 +378 0 -1.4488051 0.006411158 0.0092791227198496407 0 +379 0 -0.8092566 0.030509004 0.044700595853256767 0 +380 0 -0.8886014 0.02520137 0.036823870048860327 0 +381 1 2.5778012 0.9928464 0.010357511828567196 1 +382 0 -0.7607838 0.03426868 0.050306229433966405 0 +383 0 -0.9785019 0.020271322 0.029545823977617132 0 +384 0 -0.9785019 0.020271322 0.029545823977617132 0 +385 0 -0.557401 0.05547466 0.082338594684536998 0 +386 1 2.3652132 0.9879464 0.017495335855110976 1 +387 0 -0.6949801 0.040093575 0.059034320612404503 0 +388 0 -1.0192207 0.018361758 0.026736641899937463 0 +389 0 -1.1409771 0.013645388 0.019821680809018326 0 +390 0 -1.124477 0.014206687 0.020642900161266915 0 +391 1 3.3155096 0.9988429 0.0016703161526321778 1 +392 0 -1.2114133 0.011485413 0.016665839255833863 0 +393 0 -0.5897026 0.051427457 0.07616998626950558 0 +394 0 -0.640959 0.045573767 0.067294398351800941 0 +395 0 -1.2114133 0.011485413 0.016665839255833863 0 +396 0 -1.3728285 0.007728782 0.011193587824129689 0 +397 0 -1.1095692 0.014733352 0.021413872833133944 0 +398 0 -0.7600371 0.034329955 0.050397768234961074 0 +399 0 -0.51268965 0.06157309 0.091683707897468658 0 +400 1 2.7485878 0.9953029 0.0067924252242985176 1 +401 0 -1.0517125 0.016965643 0.024686254694263519 0 +402 0 -0.37533355 0.08442652 0.1272524232561946 0 +403 0 -0.38031232 0.08347789 0.12575841352883249 0 +404 0 -0.60584116 0.04951152 0.073258950956581831 0 +405 0 -1.0646522 0.016439164 0.023913804967856618 0 +406 0 -0.8254003 0.029347915 0.042973818276762166 0 +407 0 -1.0646522 0.016439164 0.023913804967856618 0 +408 0 -0.3370325 0.092056885 0.13932618371482772 0 +409 0 -1.117744 0.014442215 0.020987632993847882 0 +410 0 -1.0646522 0.016439164 0.023913804967856618 0 +411 0 ? ? ? 0 +412 1 2.13703 0.97897756 0.030652302483661339 1 +413 0 -1.3156877 0.00889374 0.012888353285340614 0 +414 1 2.5419102 0.99218637 0.011316959928801083 1 +415 0 0.512103 0.45388854 0.87273265130664157 1 +416 1 2.0137136 0.9716782 0.04144949701081696 1 +417 0 -1.0646522 0.016439164 0.023913804967856618 0 +418 0 -0.5104429 0.061895512 0.092179472935704498 0 +419 0 -1.3045864 0.009139484 0.013246111887140222 0 +420 0 -0.38752788 0.08212027 0.12362296481221816 0 +421 1 2.6672237 0.9942599 0.0083050815067916648 1 +422 0 -0.12513776 0.14631653 0.22822684762531151 0 +423 0 -1.0483546 0.017104952 0.024890718331546519 0 +424 0 -1.0517125 0.016965643 0.024686254694263519 0 +425 1 3.3923504 0.9990433 0.001380907436723376 1 +426 0 0.09367403 0.22764376 0.37266166457386768 1 +427 1 1.9899356 0.9700111 0.043926816071625768 1 +428 0 -1.0646522 0.016439164 0.023913804967856618 0 +429 0 -1.1461768 0.0134730805 0.019569676334993513 0 +430 0 -0.41650113 0.07686924 0.11539308004870612 0 +431 0 -1.928626 0.0019615057 0.0028326336250191013 0 +432 0 -1.1261332 0.014149335 0.020558969063149787 0 +433 0 -0.7042185 0.039221883 0.057724802498689075 0 +434 0 2.51265 0.99160385 6.8960565701084429 1 +435 1 2.8290646 0.9961487 0.0055669704170238895 1 +436 1 2.0754092 0.97559476 0.035646087064005946 1 +437 0 -1.1095692 0.014733352 0.021413872833133944 0 +438 0 -0.926998 0.02296697 0.033520759141558984 0 +439 0 -1.1535558 0.013232234 0.019217505816434433 0 +440 1 1.6831009 0.93798006 0.0923708477238116 1 +441 0 0.09868149 0.22983244 0.37675573895529579 1 +442 0 -0.5905317 0.05132734 0.076017724923000532 0 +443 0 -0.7269715 0.037151884 0.054619856517827158 0 +444 0 -1.7038846 0.0034180398 0.0049396357692610034 0 +445 0 -0.9785019 0.020271322 0.029545823977617132 0 +446 0 -0.8886014 0.02520137 0.036823870048860327 0 +447 0 -1.1535558 0.013232234 0.019217505816434433 0 +448 0 -0.5897026 0.051427457 0.07616998626950558 0 +449 1 3.2989779 0.9987946 0.0017400515007660342 1 +450 0 -1.0155473 0.018526526 0.026978818217978866 0 +451 0 -1.1535558 0.013232234 0.019217505816434433 0 +452 0 -1.1308969 0.01398564 0.020319437719197881 0 +453 1 2.3809402 0.98840165 0.016830673820750357 1 +454 0 -0.72942686 0.036934886 0.054294751351442259 0 +455 1 0.5420487 0.47233364 1.0821218062099403 1 +456 1 2.5631464 0.9925839 0.010739083310264113 1 +457 1 2.2057912 0.98221195 0.025893723576704865 1 +458 0 -1.043272 0.017317954 0.025203398068479654 0 +459 0 -0.92309535 0.023184946 0.033842660284377046 0 +460 0 -1.1986376 0.011850395 0.017198614103561393 0 +461 0 -0.1373853 0.14256682 0.22190384045518899 0 +462 0 -1.2156788 0.011366043 0.016491635093705537 0 +463 0 -0.9639507 0.020999847 0.03061900956174958 0 +464 0 -1.1095692 0.014733352 0.021413872833133944 0 +465 1 2.4820158 0.9909478 0.013119056851342249 1 +466 1 2.5902734 0.9930626 0.010043408280474288 1 +467 1 2.6717882 0.9943241 0.0082119371806173413 1 +468 0 -1.1095692 0.014733352 0.021413872833133944 0 +469 0 -1.0053848 0.018989945 0.027660171012904522 0 +470 0 -0.9095942 0.023954712 0.034980005022197523 0 +471 0 -1.2156788 0.011366043 0.016491635093705537 0 +472 0 -0.9017951 0.024410678 0.035654127331752536 0 +473 0 -1.1095692 0.014733352 0.021413872833133944 0 +474 0 -1.1535558 0.013232234 0.019217505816434433 0 +475 0 -1.0517125 0.016965643 0.024686254694263519 0 +476 0 -1.0124358 0.018667221 0.027185644374675546 0 +477 0 -1.1095692 0.014733352 0.021413872833133944 0 +478 0 -0.84103805 0.028264133 0.041363875221673722 0 +479 1 3.1646245 0.9983193 0.0024267372497735222 1 +480 0 -0.8665246 0.026580548 0.038866490244506509 0 +481 0 -0.37773794 0.08396719 0.12652881642258054 0 +482 1 2.4817219 0.9909412 0.013128602336232942 1 +483 1 3.16094 0.99830395 0.0024489605326631172 1 +484 0 -1.043272 0.017317954 0.025203398068479654 0 +485 0 -0.32016066 0.095610835 0.14498438630213031 0 +486 0 -0.9095942 0.023954712 0.034980005022197523 0 +487 1 3.006631 0.9975161 0.0035879744444351036 1 +488 1 1.4860514 0.90274173 0.14761479581436426 1 +489 1 -0.07631941 0.16207944 2.6252270130306306 0 +490 0 -0.8886014 0.02520137 0.036823870048860327 0 +491 1 2.8575056 0.99640983 0.0051888354799423742 1 +492 0 -0.9630173 0.021047443 0.030689150949999997 0 +493 1 3.4752972 0.9992209 0.0011244306892646959 1 +494 0 -0.3638718 0.08664767 0.13075659609034265 0 +495 0 -0.9095942 0.023954712 0.034980005022197523 0 +496 0 -0.5897026 0.051427457 0.07616998626950558 0 +497 0 -0.8154381 0.030059254 0.044031479309337471 0 +498 0 -1.2618061 0.01015102 0.014719663419588126 0 +499 0 -1.2618061 0.01015102 0.014719663419588126 0 +500 0 -1.4653347 0.0061554867 0.0089079346640132586 0 +501 0 -1.2618061 0.01015102 0.014719663419588126 0 +502 0 -1.292458 0.009415659 0.013648279806877021 0 +503 0 -1.4161462 0.0069477297 0.010058437491048256 0 +504 0 -0.8886014 0.02520137 0.036823870048860327 0 +505 0 -0.4987291 0.06360224 0.094806609865484295 0 +506 1 2.7913313 0.9957729 0.0061113445653601627 1 +507 0 -0.32945326 0.09363849 0.14184149411512567 0 +508 0 -1.1535558 0.013232234 0.019217505816434433 0 +509 0 -0.9785019 0.020271322 0.029545823977617132 0 +510 0 -0.8886014 0.02520137 0.036823870048860327 0 +511 0 -1.2988322 0.009269498 0.01343542569044209 0 +512 0 -1.1535558 0.013232234 0.019217505816434433 0 +513 0 -0.9095942 0.023954712 0.034980005022197523 0 +514 1 3.3379297 0.99890536 0.0015800956668196994 1 +515 1 3.5692976 0.9993827 0.0008908871693913803 1 +516 0 -0.5897026 0.051427457 0.07616998626950558 0 +517 0 -0.7576831 0.03452383 0.050687443773780996 0 +518 0 -1.0003859 0.019222053 0.02800155445627163 0 +519 1 2.4472048 0.99014044 0.014294928329511082 1 +520 0 -1.204251 0.011688637 0.016962466394122074 0 +521 0 -1.4178827 0.0069181086 0.010015404819307995 0 +522 1 1.1353608 0.79562986 0.32983067520069587 1 +523 1 2.4855974 0.991027 0.013003735029207939 1 +524 0 -1.2114133 0.011485413 0.016665839255833863 0 +525 0 -1.0939798 0.015304665 0.022250671629517431 0 +526 0 -1.1095692 0.014733352 0.021413872833133944 0 +527 0 -1.4161462 0.0069477297 0.010058437491048256 0 +528 0 -0.93557656 0.022494845 0.032823784344435981 0 +529 0 -0.9630173 0.021047443 0.030689150949999997 0 +530 1 2.4382513 0.9899215 0.014613954852764939 1 +531 0 -0.8254003 0.029347915 0.042973818276762166 0 +532 0 -0.9263915 0.023000715 0.033570588948334805 0 +533 0 -1.2114133 0.011485413 0.016665839255833863 0 +534 0 -1.1461768 0.0134730805 0.019569676334993513 0 +535 0 -1.113402 0.014596137 0.021212967863552434 0 +536 0 -1.4465218 0.006447295 0.0093315950043344458 0 +537 0 -1.3156877 0.00889374 0.012888353285340614 0 +538 0 -1.2618061 0.01015102 0.014719663419588126 0 +539 0 -1.3239301 0.008715526 0.012628960739826556 0 +540 0 -1.0392807 0.017487051 0.025451673383951559 0 +541 0 -1.0517125 0.016965643 0.024686254694263519 0 +542 0 -0.53340137 0.058673874 0.087233457472134818 0 +543 0 -1.2618061 0.01015102 0.014719663419588126 0 +544 0 -1.0018501 0.01915378 0.02790112982807072 0 +545 0 -1.2988322 0.009269498 0.01343542569044209 0 +546 1 4.0447593 0.99980986 0.00027433841014312443 1 +547 0 -0.68175036 0.04137425 0.060960403522021971 0 +548 0 -0.757901 0.034505837 0.050660557155260993 0 +549 1 2.1623328 0.98022974 0.028808183166648332 1 +550 0 -1.2114133 0.011485413 0.016665839255833863 0 +551 0 -1.2416301 0.010665795 0.015470137413992897 0 +552 0 -0.8777449 0.025870657 0.037814751018648084 0 +553 0 0.71006846 0.57578754 1.2371411129056633 1 +554 0 -1.0517125 0.016965643 0.024686254694263519 0 +555 0 0.34858382 0.356611 0.6362368544695326 1 +556 0 -0.5773324 0.05294328 0.078477265209329058 0 +557 0 -1.1986376 0.011850395 0.017198614103561393 0 +558 0 -1.1461768 0.0134730805 0.019569676334993513 0 +559 0 -1.2988322 0.009269498 0.01343542569044209 0 +560 0 -1.4465218 0.006447295 0.0093315950043344458 0 +561 0 -1.4465218 0.006447295 0.0093315950043344458 0 +562 0 -1.2416301 0.010665795 0.015470137413992897 0 +563 0 -1.2114133 0.011485413 0.016665839255833863 0 +564 0 -1.242877 0.010633245 0.01542267261125157 0 +565 1 3.496207 0.9992602 0.0010677193362146344 1 +566 0 -1.2679701 0.009998713 0.014497693739111505 0 +567 0 -0.76216024 0.034156 0.050137904027057756 0 +568 1 1.8593472 0.9590243 0.060360708597549641 1 +569 1 2.257011 0.98429865 0.022831984218390725 1 +570 1 2.2973976 0.98577243 0.020673461174525863 1 +571 1 3.3566217 0.99895483 0.0015086464229099373 1 +572 0 -1.2114133 0.011485413 0.016665839255833863 0 +573 0 -1.0646522 0.016439164 0.023913804967856618 0 +574 1 1.9174787 0.9643245 0.052409431057596163 1 +575 0 -1.3156877 0.00889374 0.012888353285340614 0 +576 0 -1.2988322 0.009269498 0.01343542569044209 0 +577 0 -1.0646522 0.016439164 0.023913804967856618 0 +578 0 -1.0646522 0.016439164 0.023913804967856618 0 +579 0 -1.2416301 0.010665795 0.015470137413992897 0 +580 0 -1.1781884 0.012458586 0.018086844101384787 0 +581 1 3.0078323 0.9975234 0.003577371212861954 1 +582 1 3.5221264 0.9993062 0.0010012864134882542 1 +583 0 -1.0517125 0.016965643 0.024686254694263519 0 +584 0 -1.5559932 0.00492325 0.007120290349874615 0 +585 0 -0.8886014 0.02520137 0.036823870048860327 0 +586 1 3.7748106 0.9996289 0.00053548135645133934 1 +587 0 -1.1261332 0.014149335 0.020558969063149787 0 +588 1 1.513865 0.9086263 0.13824100183440888 1 +589 0 -1.1535558 0.013232234 0.019217505816434433 0 +590 1 1.4137613 0.8858426 0.17487768233671636 1 +591 1 2.6803715 0.9944428 0.008039674941419564 1 +592 1 1.7123015 0.9420577 0.086112624800305834 1 +593 0 -1.043272 0.017317954 0.025203398068479654 0 +594 1 2.4865408 0.9910478 0.012973452647715817 1 +595 0 -1.2988322 0.009269498 0.01343542569044209 0 +596 0 -1.1537459 0.013226084 0.019208514989069607 0 +597 0 -1.1762671 0.012517288 0.01817260443408111 0 +598 0 -1.2114133 0.011485413 0.016665839255833863 0 +599 0 -0.23754154 0.11483499 0.1759816659252687 0 +600 0 -1.2114133 0.011485413 0.016665839255833863 0 +601 0 -0.7576831 0.03452383 0.050687443773780996 0 +602 0 -1.2618061 0.01015102 0.014719663419588126 0 +603 1 1.2366467 0.83343154 0.26286439380293886 1 +604 1 1.1037434 0.78259724 0.35365806563955249 1 +605 1 2.4797065 0.9908963 0.013194033996522414 1 +606 0 -1.1425668 0.013592481 0.019744297537447716 0 +607 0 -0.8886014 0.02520137 0.036823870048860327 0 +608 1 2.682395 0.9944705 0.0079995525539330661 1 +609 0 -1.1535558 0.013232234 0.019217505816434433 0 +610 1 1.8788853 0.96088487 0.057564513764289997 1 +611 1 2.2919562 0.9855821 0.020952021217181316 1 +612 1 3.6261826 0.9994638 0.00077378543531000823 1 +613 0 -0.5888309 0.051532917 0.076330389378713615 0 +614 0 -0.84157866 0.028227368 0.041309292980877294 0 +615 0 -1.0343657 0.017697506 0.025760732370776564 0 +616 0 -1.2114133 0.011485413 0.016665839255833863 0 +617 0 ? ? ? 0 +618 0 -1.2618061 0.01015102 0.014719663419588126 0 +619 0 -1.2988322 0.009269498 0.01343542569044209 0 +620 0 -1.2114133 0.011485413 0.016665839255833863 0 +621 0 -0.6180498 0.04810729 0.071129119789409653 0 +622 0 -1.1982852 0.011860623 0.01721354671122631 0 +623 0 -0.8886014 0.02520137 0.036823870048860327 0 +624 0 -1.0238441 0.018156424 0.026434897580118549 0 +625 0 -0.37955323 0.0836219 0.12598510895000004 0 +626 1 1.7627405 0.948514 0.076259050317554145 1 +627 0 -0.12628824 0.14596084 0.2276258680145801 0 +628 0 -0.9785019 0.020271322 0.029545823977617132 0 +629 0 -1.1095692 0.014733352 0.021413872833133944 0 +630 0 -0.59879136 0.05033999 0.074516990666586291 0 +631 0 -1.2988322 0.009269498 0.01343542569044209 0 +632 0 -0.8886014 0.02520137 0.036823870048860327 0 +633 1 1.7914765 0.9518819 0.071145527377270423 1 +634 0 -1.0517125 0.016965643 0.024686254694263519 0 +635 0 -0.76496696 0.033927336 0.049796388676926706 0 +636 1 1.9455588 0.9666421 0.048946292556533537 1 +637 0 -0.2736899 0.10604059 0.16171876640514166 0 +638 0 -1.1095692 0.014733352 0.021413872833133944 0 +639 0 -1.1986376 0.011850395 0.017198614103561393 0 +640 0 -1.1708342 0.012684766 0.018417307445062352 0 +641 0 -1.2114133 0.011485413 0.016665839255833863 0 +642 0 -1.2114133 0.011485413 0.016665839255833863 0 +643 0 -0.8886014 0.02520137 0.036823870048860327 0 +644 0 -0.9785019 0.020271322 0.029545823977617132 0 +645 0 -1.2114133 0.011485413 0.016665839255833863 0 +646 0 -0.59787124 0.050449084 0.074682734234965215 0 +647 0 -1.0088996 0.018828396 0.027422612934787257 0 +648 1 2.038984 0.97335124 0.038967589989537407 1 +649 0 -1.2114133 0.011485413 0.016665839255833863 0 +650 0 -0.64436394 0.045208234 0.066741970542076506 0 +651 0 -0.8884886 0.025208237 0.036834034043365216 0 +652 0 -1.1261332 0.014149335 0.020558969063149787 0 +653 0 -1.2618061 0.01015102 0.014719663419588126 0 +654 0 -1.3728285 0.007728782 0.011193587824129689 0 +655 0 -1.2114133 0.011485413 0.016665839255833863 0 +656 0 -1.2988322 0.009269498 0.01343542569044209 0 +657 0 -0.18411672 0.12899058 0.19923976600711193 0 +658 1 2.920357 0.99692595 0.0044417471973557526 1 +659 0 -0.8886014 0.02520137 0.036823870048860327 0 +660 0 -1.0646522 0.016439164 0.023913804967856618 0 +661 0 -1.4161462 0.0069477297 0.010058437491048256 0 +662 0 -0.9277877 0.022923106 0.03345599149152359 0 +663 0 -0.9277877 0.022923106 0.03345599149152359 0 +664 0 -1.2854787 0.0095783165 0.013885194738261197 0 +665 0 -0.8886014 0.02520137 0.036823870048860327 0 +666 0 -0.7735714 0.033235505 0.048763604975984247 0 +667 0 -1.3728285 0.007728782 0.011193587824129689 0 +668 1 1.1245754 0.79125047 0.33779364831252268 1 +669 1 3.138155 0.9982056 0.0025910942749578212 1 +670 1 2.8930492 0.9967115 0.0047521314061252529 1 +671 0 -1.0711483 0.016180947 0.023535100415445195 0 +672 0 -1.3151383 0.008905744 0.012905826640471321 0 +673 0 -0.6507327 0.044532 0.065720540349053072 0 +674 0 -1.0646522 0.016439164 0.023913804967856618 0 +675 0 -0.7845539 0.03237224 0.047475935870418158 0 +676 0 -1.0053848 0.018989945 0.027660171012904522 0 +677 0 -1.1535558 0.013232234 0.019217505816434433 0 +678 0 -0.8886014 0.02520137 0.036823870048860327 0 +679 0 -0.9785019 0.020271322 0.029545823977617132 0 +680 1 3.7458398 0.9996013 0.0005753106703117719 1 +681 1 3.7473078 0.99960274 0.00057324605682610265 1 +682 0 -1.2606217 0.01018055 0.014762702933716242 0 +683 0 -0.8886014 0.02520137 0.036823870048860327 0 +684 0 -0.8886014 0.02520137 0.036823870048860327 0 +685 0 -0.8886014 0.02520137 0.036823870048860327 0 +686 0 -0.8886014 0.02520137 0.036823870048860327 0 +687 0 -0.9116796 0.023834204 0.034801893494115908 0 +688 0 -1.1095692 0.014733352 0.021413872833133944 0 +689 0 -1.5070968 0.005553787 0.0080347534846788634 0 +690 0 -1.0088996 0.018828396 0.027422612934787257 0 +691 1 2.6980603 0.99467987 0.0076958173153511764 1 +692 0 -1.0517125 0.016965643 0.024686254694263519 0 +693 0 -1.2912332 0.009444007 0.01368956606633783 0 +694 0 -1.1383268 0.013734051 0.01995136965321119 0 +695 0 -0.9785019 0.020271322 0.029545823977617132 0 +696 1 2.4236615 0.9895544 0.015149066773324307 1 +697 1 2.013682 0.97167605 0.041452682933010118 1 +698 1 2.0937462 0.9766534 0.034081437533218267 1 diff --git a/test/BaselineOutput/Common/LogisticRegression/net9.0/LogisticRegression-non-negative-CV-breast-cancer.txt b/test/BaselineOutput/Common/LogisticRegression/net9.0/LogisticRegression-non-negative-CV-breast-cancer.txt new file mode 100644 index 0000000000..fa7532ad85 --- /dev/null +++ b/test/BaselineOutput/Common/LogisticRegression/net9.0/LogisticRegression-non-negative-CV-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +5 1 12.633929 0.9999967 4.7295306510363196E-06 1 +6 0 -1.3843784 0.20030673 0.32248134553051777 0 +8 0 -4.7242136 0.0087995725 0.012751285145230893 0 +9 0 -5.136591 0.0058433474 0.0084548948942712102 0 +10 0 -5.6091967 0.0036506357 0.0052763910016335762 0 +11 0 -5.8695636 0.0028161525 0.0040685808485482122 0 +18 1 7.2667055 0.9993021 0.0010072239466017226 1 +20 1 7.564786 0.99948186 0.00074771632076294251 1 +21 1 7.389447 0.9993826 0.00089097321383678973 1 +25 1 1.348588 0.7938987 0.332973157953321 1 +28 0 -5.8695636 0.0028161525 0.0040685808485482122 0 +31 0 -5.1808686 0.005591677 0.0080897234281375775 0 +32 1 7.291772 0.9993193 0.00098235531165322479 1 +35 0 -5.8695636 0.0028161525 0.0040685808485482122 0 +37 0 -1.0236454 0.26431793 0.44284566387899665 0 +40 0 ? ? ? 0 +41 1 2.7849321 0.94185615 0.086421367856595113 1 +44 1 7.8208895 0.9995989 0.00057875169935374328 1 +45 0 -5.776985 0.003088473 0.0044626197323314855 0 +46 1 4.362952 0.98741955 0.018264891844072241 1 +48 0 -3.8099475 0.02166938 0.031605999137608386 0 +50 1 2.3504057 0.9129665 0.13136618746901846 1 +51 1 -0.38573265 0.404745 1.3049147918875741 0 +52 1 4.7615347 0.99152005 0.012286152805355393 1 +54 1 7.369527 0.9993702 0.00090887057002802098 1 +56 1 4.5134735 0.9891585 0.015726365669399833 1 +60 1 2.0933895 0.890259 0.16770293488464477 1 +63 1 -0.43272972 0.3934747 1.3456572270003335 0 +64 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +66 0 -4.4490027 0.011555137 0.016767602638168632 0 +68 1 12.4205675 0.99999595 5.8474219794828371E-06 1 +69 0 -5.372825 0.004619562 0.0066800606698008674 0 +70 0 -3.9446263 0.018990817 0.02766145298110308 0 +71 1 8.003949 0.999666 0.00048197588806008033 1 +72 0 -1.4691668 0.1870693 0.29879571732730276 0 +73 1 8.232537 0.9997342 0.00038348628329455832 1 +74 1 1.5337038 0.82254755 0.28182900529686483 1 +76 0 -4.9947886 0.0067275856 0.0097386499935240151 0 +77 0 -4.450024 0.011543477 0.016750584032023265 0 +79 0 -5.807169 0.0029969194 0.0043301325946410044 0 +82 0 -4.0819635 0.016594283 0.024141353054393848 0 +88 0 -4.4490027 0.011555137 0.016767602638168632 0 +90 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +91 0 -5.872798 0.0028070842 0.0040554611870710329 0 +92 0 -4.4490027 0.011555137 0.016767602638168632 0 +93 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +95 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +96 0 -6.238731 0.0019485275 0.0028138733449574417 0 +97 0 -4.08307 0.01657624 0.02411488285262782 0 +98 1 8.65561 0.9998259 0.00025120253004900898 1 +99 1 11.468547 0.99998957 1.5048560434990871E-05 1 +100 1 3.7765074 0.97761023 0.03266871220997121 1 +102 0 -4.0354333 0.017370936 0.025281182739844577 0 +104 1 12.724471 0.999997 4.2995726784359523E-06 1 +105 1 2.762721 0.94062775 0.088304194559905208 1 +106 1 8.566487 0.9998097 0.00027459643320268699 1 +108 0 -5.887544 0.0027661077 0.0039961793075080502 0 +109 1 7.138543 0.9992067 0.0011449127273942801 1 +111 1 2.7871408 0.94197696 0.086236314928145763 1 +112 1 9.633872 0.99993455 9.4421565109537245E-05 1 +113 1 9.297807 0.9999084 0.00013217472169822292 1 +115 0 -5.0742254 0.006217037 0.0089972855480379404 0 +117 1 11.032926 0.99998385 2.3303837404214046E-05 1 +120 0 -4.8137045 0.008052365 0.011664132147019496 0 +121 0 -3.985405 0.01824582 0.02656625937708185 0 +122 1 9.102151 0.9998886 0.00016072673996382048 1 +123 1 3.835556 0.97886693 0.030815338823386065 1 +125 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +128 1 3.9304113 0.9807425 0.028053676426704734 1 +129 0 -6.4250317 0.0016178603 0.0023359691512568463 0 +131 0 -5.1808686 0.005591677 0.0080897234281375775 0 +132 1 8.0213175 0.9996717 0.00047371798610370159 1 +133 0 -5.3296547 0.004822375 0.0069740454692729815 0 +137 0 -6.1923256 0.0020408914 0.0029473926413165152 0 +138 0 -4.7241282 0.0088003175 0.012752369579763029 0 +141 0 -6.5582585 0.0014163447 0.0020448018331478031 0 +144 0 -5.8695636 0.0028161525 0.0040685808485482122 0 +145 0 ? ? ? 0 +147 0 -5.7106104 0.0032997269 0.0047683710890380751 0 +150 0 -5.6091967 0.0036506357 0.0052763910016335762 0 +151 1 5.1189547 0.9940533 0.008604879280406149 1 +152 1 9.233193 0.99990225 0.00014103266690546063 1 +154 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +156 0 -5.820821 0.0029564041 0.0042715068256996428 0 +161 0 -4.2750263 0.013720803 0.019931990714553656 0 +164 0 ? ? ? 0 +167 1 8.728393 0.9998381 0.00023357134637246152 1 +169 0 -6.667059 0.0012705184 0.0018341360014305023 0 +171 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +173 1 16.565205 0.99999994 8.5991327994145617E-08 1 +174 1 5.9240236 0.9973327 0.003853253006618053 1 +176 0 -5.1808686 0.005591677 0.0080897234281375775 0 +177 1 4.5226355 0.9892563 0.015583714327774979 1 +179 1 1.7973394 0.85782474 0.22124516572731823 1 +180 0 -5.6091967 0.0036506357 0.0052763910016335762 0 +181 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +183 1 9.378575 0.9999155 0.00012194085271734281 1 +187 1 15.031856 0.9999997 4.2995669122556443E-07 1 +188 1 9.1637125 0.9998952 0.00015118067097744242 1 +189 0 -4.7185564 0.008849053 0.012823305568612981 0 +191 1 11.08418 0.9999846 2.2185932549356626E-05 1 +192 0 -4.7717648 0.008394366 0.012161626676841389 0 +196 0 6.86773 0.99896026 9.909556724918394 1 +198 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +199 0 -5.5036306 0.004055447 0.0058626691195546371 0 +201 1 9.712966 0.9999395 8.7283835619421935E-05 1 +202 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +204 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +205 1 12.771848 0.99999714 4.1275895252744015E-06 1 +206 1 4.562008 0.9896668 0.014985184496894443 1 +207 0 -5.6091967 0.0036506357 0.0052763910016335762 0 +209 0 -4.145465 0.0155892 0.022667609096222847 0 +210 1 15.273984 0.99999976 3.4396534272948301E-07 1 +211 1 9.317233 0.9999101 0.00012968074693731126 1 +212 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +216 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +218 1 7.7962294 0.9995889 0.00059320411095478676 1 +219 0 -2.8899984 0.052650195 0.078030861051999856 0 +223 1 5.9366674 0.99736613 0.0038048836658159005 1 +226 1 10.313518 0.9999668 4.7897963370785976E-05 1 +228 0 -5.6091967 0.0036506357 0.0052763910016335762 0 +233 1 5.720605 0.996733 0.0047209864523075872 1 +237 1 5.9035883 0.9972778 0.0039326650133254511 1 +239 1 5.305252 0.9950591 0.0071459192280271143 1 +240 0 -2.4951277 0.076200455 0.11434825993951341 0 +241 0 -4.2382755 0.014227127 0.02067281374647327 0 +242 0 -5.1808686 0.005591677 0.0080897234281375775 0 +244 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +246 1 11.305726 0.9999877 1.7714321792245208E-05 1 +247 1 2.3066397 0.90942544 0.13697273675794658 1 +248 0 -3.579121 0.027142916 0.039700210916466902 0 +249 0 ? ? ? 0 +250 0 -6.509516 0.0014869863 0.0021468643262315163 0 +252 0 4.2280703 0.985629 6.1206979559151264 1 +254 1 5.197297 0.9944989 0.0079583072109629376 1 +257 0 -5.5036306 0.004055447 0.0058626691195546371 0 +258 0 -4.8149357 0.008042537 0.011649837951791604 0 +259 0 2.3515387 0.91305643 3.5237769073872456 1 +260 1 10.040982 0.9999564 6.2861028359088115E-05 1 +262 1 9.611833 0.99993306 9.6571490550947405E-05 1 +267 1 4.1406794 0.9843372 0.022775461436525875 1 +268 1 10.115634 0.9999596 5.8303296725358504E-05 1 +269 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +271 0 -4.08307 0.01657624 0.02411488285262782 0 +272 1 4.1406794 0.9843372 0.022775461436525875 1 +275 0 ? ? ? 0 +276 0 -5.5036306 0.004055447 0.0058626691195546371 0 +277 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +278 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +279 1 7.266366 0.99930185 0.0010075681521692589 1 +280 0 -4.8149357 0.008042537 0.011649837951791604 0 +283 1 5.7055264 0.99668354 0.0047925949680514479 1 +284 1 7.0209074 0.9991078 0.001287778728059478 1 +285 1 14.941096 0.9999997 4.2995669122556443E-07 1 +288 1 2.027934 0.88369894 0.17837314082060335 1 +290 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +291 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +293 1 4.40994 0.9879901 0.017431536595943067 1 +296 0 1.5062141 0.81849945 2.4619541424010967 1 +297 0 ? ? ? 0 +299 1 7.9362307 0.99964255 0.00051578216702891913 1 +300 1 7.8511133 0.99961084 0.00056154663621664426 1 +301 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +303 0 -5.5468016 0.0038847607 0.0056154393381534417 0 +304 1 5.3866405 0.9954435 0.006588628765737457 1 +308 1 7.8222094 0.9995994 0.000578063492888762 1 +309 0 -1.5875807 0.16972455 0.26833806094724499 0 +311 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +312 1 2.3317823 0.91147524 0.1337246266710603 1 +314 0 -6.2978916 0.0018368005 0.0026523796391730529 0 +316 1 3.281417 0.96378577 0.053215598259643997 1 +317 1 9.331021 0.99991137 0.00012787476790525535 1 +319 0 2.955886 0.95054096 4.3376219476753999 1 +321 0 ? ? ? 0 +323 1 5.714961 0.99671453 0.0047477313857159737 1 +327 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +328 1 1.3741941 0.79805696 0.32543637459307595 1 +329 1 7.8807306 0.99962217 0.00054520201628348698 1 +331 0 -3.6245584 0.025968524 0.037959700430001471 0 +332 0 -3.8635855 0.020560969 0.029972405285289192 0 +333 1 4.9095945 0.9926785 0.010601515374353389 1 +336 1 4.9874 0.99322283 0.0098106676303476859 1 +338 0 -6.2978916 0.0018368005 0.0026523796391730529 0 +343 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +344 1 10.538409 0.9999735 3.8266647312775575E-05 1 +346 0 -4.3159094 0.01317841 0.019138815170622079 0 +347 0 -6.638634 0.0013071029 0.0018869843120168435 0 +348 1 -0.16002369 0.46007922 1.1200457887844508 0 +349 1 4.581849 0.98986775 0.014692310844729243 1 +350 0 -4.821405 0.0079910895 0.011575015507850202 0 +352 0 0.27724934 0.56887174 1.2138109505187624 1 +353 1 9.354077 0.9999134 0.00012495080664652852 1 +354 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +355 0 -4.9013267 0.007381814 0.010689208138700793 0 +358 1 4.2513685 0.98595536 0.020405769680748389 1 +360 1 16.688137 0.99999994 8.5991327994145617E-08 1 +361 1 6.0628357 0.9976776 0.0033543765836277469 1 +366 1 14.49912 0.9999995 6.8793076746672365E-07 1 +368 0 -6.097053 0.0022444383 0.0032416792094242604 0 +370 0 -4.65952 0.009382147 0.013599473804873082 0 +371 0 -6.097053 0.0022444383 0.0032416792094242604 0 +373 0 -4.452237 0.011518254 0.016713770428900425 0 +376 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +377 0 -6.2978916 0.0018368005 0.0026523796391730529 0 +378 0 -4.3086767 0.0132728005 0.019276816997533676 0 +379 0 -2.7153587 0.062073134 0.092452660428008815 0 +381 1 9.812264 0.9999452 7.9028192541775766E-05 1 +383 0 -6.5582585 0.0014163447 0.0020448018331478031 0 +384 0 -6.5582585 0.0014163447 0.0020448018331478031 0 +387 0 -2.8717957 0.053565547 0.079425501709729798 0 +388 0 -5.695587 0.0033495063 0.0048404271767004611 0 +389 0 -4.072902 0.016742803 0.024359254031804434 0 +391 1 9.565747 0.9999299 0.00010112934308253928 1 +392 0 -5.5036306 0.004055447 0.0058626691195546371 0 +395 0 -5.5036306 0.004055447 0.0058626691195546371 0 +396 0 -4.8149357 0.008042537 0.011649837951791604 0 +398 0 -5.328548 0.0048276894 0.0069817499796982112 0 +399 0 -6.0959463 0.0022469182 0.0032452649712945937 0 +404 0 -6.110054 0.0022155114 0.0031998532735132527 0 +406 0 -4.7230215 0.008809976 0.012766427991708861 0 +409 0 -5.090061 0.0061199586 0.0088563619203968699 0 +413 0 -3.6695004 0.024855651 0.036312300406847971 0 +414 1 6.567313 0.99859643 0.0020263461236442811 1 +415 0 -1.1182337 0.24633907 0.40801248582177035 0 +416 1 8.784802 0.999847 0.00022075662126012742 1 +418 0 -2.2538877 0.095014654 0.14403366354066907 0 +419 0 -6.3688607 0.0017111779 0.0024708224424948369 0 +422 0 -3.271699 0.036554947 0.053725706248787794 0 +423 0 -3.9446263 0.018990817 0.02766145298110308 0 +428 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +429 0 -5.8695636 0.0028161525 0.0040685808485482122 0 +430 0 -6.059307 0.0023305747 0.0033662326764313968 0 +434 0 4.338537 0.98711264 6.2778995840047314 1 +436 1 2.2573948 0.9052865 0.1435536688592095 1 +439 0 -5.4604597 0.004233601 0.0061207608129045408 0 +440 1 10.56088 0.9999741 3.7406711505747454E-05 1 +441 0 -2.699727 0.062989466 0.093862827771854257 0 +442 0 -5.3601537 0.0046781953 0.006765045690427708 0 +449 1 10.89008 0.99998134 2.691553593234519E-05 1 +450 0 -4.561038 0.010343106 0.014999654331028725 0 +451 0 -5.4604597 0.004233601 0.0061207608129045408 0 +452 0 -5.553271 0.003859807 0.0055792989222687413 0 +453 1 9.723119 0.9999401 8.6423870594618947E-05 1 +454 0 -6.4936504 0.00151073 0.0021811706934101414 0 +455 1 -0.44515133 0.3905142 1.356553104435781 0 +456 1 10.306062 0.99996656 4.8241940133472131E-05 1 +457 1 9.489881 0.9999244 0.00010904112129145599 1 +464 0 -5.8263927 0.0029400256 0.0042478078161316008 0 +465 1 9.442241 0.9999207 0.00011437299629627494 1 +466 1 9.602742 0.99993247 9.7431461624590437E-05 1 +467 1 8.004737 0.9996662 0.0004816318078681825 1 +474 0 -5.4604597 0.004233601 0.0061207608129045408 0 +480 0 -5.522855 0.0039785365 0.0057512632336553576 0 +482 1 16.850868 0.99999994 8.5991327994145617E-08 1 +483 1 11.242028 0.9999869 1.8918215632667518E-05 1 +484 0 -5.0468903 0.006388224 0.0092458231065348524 0 +487 1 14.043946 0.9999992 1.1178876637117433E-06 1 +489 1 -1.1627903 0.23816064 2.0699930912863778 0 +492 0 -5.200093 0.0054857926 0.0079361137376705946 0 +493 1 10.157264 0.9999612 5.5981438976225112E-05 1 +495 0 -5.5660257 0.0038110747 0.0055087222594122831 0 +497 0 -5.825286 0.0029432715 0.0042525044643163725 0 +501 0 -5.1376977 0.0058369217 0.0084455701686528568 0 +502 0 -4.81817 0.008016774 0.011612370034127566 0 +504 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +507 0 -5.5367174 0.003923979 0.0056722413910982785 0 +510 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +513 0 -5.5660257 0.0038110747 0.0055087222594122831 0 +514 1 11.338149 0.9999881 1.7198367596743667E-05 1 +517 0 -6.2978916 0.0018368005 0.0026523796391730529 0 +519 1 7.5110598 0.9994533 0.00078892810756466295 1 +520 0 -6.8617964 0.0010459363 0.0015097567375380694 0 +521 0 -5.763998 0.0031287188 0.0045208630708076696 0 +522 1 5.3778677 0.9954036 0.0066465078341357451 1 +523 1 7.609372 0.99950445 0.00071510908150589211 1 +527 0 -4.4490027 0.011555137 0.016767602638168632 0 +528 0 -3.9478607 0.018930653 0.027572977935081937 0 +529 0 -5.200093 0.0054857926 0.0079361137376705946 0 +531 0 -4.7230215 0.008809976 0.012766427991708861 0 +532 0 -5.6091967 0.0036506357 0.0052763910016335762 0 +533 0 -5.5036306 0.004055447 0.0058626691195546371 0 +534 0 -5.8695636 0.0028161525 0.0040685808485482122 0 +535 0 -4.4001355 0.012126812 0.017602237761857209 0 +538 0 -5.1376977 0.0058369217 0.0084455701686528568 0 +539 0 -4.4058323 0.012058754 0.017502849915883526 0 +540 0 -4.3092737 0.013264985 0.019265389758174038 0 +541 0 -6.1923256 0.0020408914 0.0029473926413165152 0 +544 0 -4.7228975 0.008811059 0.012768004505226161 0 +546 1 11.498473 0.99998987 1.4618599387059818E-05 1 +547 0 -6.6206536 0.0013307863 0.0019211973849274147 0 +548 0 -6.2547207 0.0019176779 0.0027692805853928785 0 +549 1 5.8067255 0.99700177 0.0043320331239142381 1 +557 0 -4.821405 0.0079910895 0.011575015507850202 0 +558 0 -5.8695636 0.0028161525 0.0040685808485482122 0 +559 0 -4.7717648 0.008394366 0.012161626676841389 0 +560 0 -4.08307 0.01657624 0.02411488285262782 0 +561 0 -4.08307 0.01657624 0.02411488285262782 0 +563 0 -5.5036306 0.004055447 0.0058626691195546371 0 +565 1 13.257699 0.9999983 2.4937505927785606E-06 1 +566 0 -4.401366 0.012112076 0.01758071828983486 0 +569 1 10.79565 0.9999795 2.9581319217091579E-05 1 +577 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +578 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +581 1 9.043624 0.99988186 0.00017044487507576205 1 +582 1 7.144294 0.99921125 0.0011383722130280166 1 +584 0 -3.8004274 0.021872124 0.031905005539105288 0 +586 1 14.421212 0.99999946 7.7392213646475052E-07 1 +590 1 4.6797056 0.9908036 0.01332898532010793 1 +593 0 -5.0468903 0.006388224 0.0092458231065348524 0 +594 1 5.307637 0.9950708 0.0071288949210089256 1 +600 0 -5.5036306 0.004055447 0.0058626691195546371 0 +602 0 -5.1376977 0.0058369217 0.0084455701686528568 0 +604 1 6.821 0.99891055 0.0015726062427077678 1 +606 0 -5.506865 0.0040424042 0.0058437759739102726 0 +607 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +609 0 -5.4604597 0.004233601 0.0061207608129045408 0 +612 1 18.547813 1 -0 1 +613 0 -5.8673506 0.002822374 0.0040775819053926205 0 +614 0 -5.9319587 0.0026462588 0.0038228047458744208 0 +617 0 ? ? ? 0 +618 0 -5.1376977 0.0058369217 0.0084455701686528568 0 +619 0 -4.7717648 0.008394366 0.012161626676841389 0 +621 0 0.3422413 0.58473486 1.2678953169132128 1 +622 0 -2.9872751 0.04800406 0.070972675527967377 0 +624 0 -4.514632 0.010829078 0.015708264532154868 0 +627 0 -4.708293 0.008939526 0.012955002344643449 0 +629 0 -5.8263927 0.0029400256 0.0042478078161316008 0 +633 1 3.276002 0.9635963 0.053499264307225285 1 +634 0 -6.1923256 0.0020408914 0.0029473926413165152 0 +638 0 -5.8263927 0.0029400256 0.0042478078161316008 0 +639 0 -4.821405 0.0079910895 0.011575015507850202 0 +641 0 -5.5036306 0.004055447 0.0058626691195546371 0 +642 0 -5.5036306 0.004055447 0.0058626691195546371 0 +644 0 -6.5582585 0.0014163447 0.0020448018331478031 0 +645 0 -5.5036306 0.004055447 0.0058626691195546371 0 +649 0 -5.5036306 0.004055447 0.0058626691195546371 0 +652 0 -4.597789 0.009973612 0.014461115364606977 0 +653 0 -5.1376977 0.0058369217 0.0084455701686528568 0 +654 0 -4.8149357 0.008042537 0.011649837951791604 0 +656 0 -4.7717648 0.008394366 0.012161626676841389 0 +657 0 -1.181222 0.23483256 0.38615260125615852 0 +660 0 -6.2354965 0.0019548277 0.0028229803347829701 0 +661 0 -4.4490027 0.011555137 0.016767602638168632 0 +665 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +668 1 2.6514454 0.93410003 0.098351039887888514 1 +670 1 6.6144648 0.998661 0.0019330896360835407 1 +678 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +679 0 -6.5582585 0.0014163447 0.0020448018331478031 0 +680 1 17.841751 1 -0 1 +681 1 9.976266 0.9999535 6.7074793056570611E-05 1 +682 0 -3.9090934 0.01966424 0.028652144059552218 0 +683 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +685 0 -6.9241915 0.0009827315 0.0014184789162451506 0 +688 0 -5.8263927 0.0029400256 0.0042478078161316008 0 +689 0 -3.7146125 0.023785355 0.034729699300628886 0 +691 1 4.112502 0.9838968 0.023421106124485898 1 +692 0 -6.1923256 0.0020408914 0.0029473926413165152 0 +693 0 -5.0451193 0.006399475 0.0092621590908444542 0 +694 0 -5.6674395 0.0034447974 0.0049783717397341833 0 +696 1 6.992797 0.9990824 0.0013244442119411943 1 +697 1 5.725787 0.9967498 0.0046966576209477966 1 +698 1 7.049386 0.9991328 0.0012516305716642035 1 +0 0 -3.9587402 0.01872965 0.027277425706471844 0 +1 0 1.4086609 0.80355465 2.347800106563402 1 +2 0 -4.911849 0.0073051117 0.010577731442167546 0 +3 0 1.5250549 0.8212816 2.4842400160316997 1 +4 0 -4.0157437 0.017710233 0.025779425140335981 0 +7 0 -5.5262513 0.0039651 0.0057318009353258779 0 +12 1 -1.0070896 0.2675498 1.9021205754104284 0 +13 0 -5.8649573 0.0028291175 0.004087338250744154 0 +14 1 6.7670994 0.9988503 0.0016596409153587515 1 +15 1 0.7604084 0.6814424 0.5533364212356795 1 +16 0 -4.9443645 0.0070730555 0.010240520805639591 0 +17 0 -4.595051 0.010000683 0.014500565544495215 0 +19 0 -3.3224301 0.034809668 0.051114629699500667 0 +22 0 -5.580675 0.0037558598 0.005428761531925271 0 +23 1 ? ? ? 0 +24 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +26 0 -5.54845 0.0038783872 0.00560620848234826 0 +27 0 -4.308054 0.013280959 0.019288745432599251 0 +29 0 -6.5197678 0.0014718418 0.00212498310420812 0 +30 0 -5.580675 0.0037558598 0.005428761531925271 0 +33 0 -5.578018 0.0037658145 0.0054431773063631141 0 +34 0 -5.239254 0.005276226 0.0076321372386148386 0 +36 1 8.057671 0.99968344 0.00045677223115128229 1 +38 1 4.114853 0.983934 0.023366570350680384 1 +39 1 -0.31334305 0.42229894 1.2436634757191947 0 +42 1 5.392255 0.9954689 0.0065518292521607709 1 +43 1 -0.9836645 0.27216527 1.877445120365532 0 +47 0 -6.8532963 0.0010548553 0.0015226377274540548 0 +49 1 3.7823238 0.9777372 0.032481367984933111 1 +53 1 4.716364 0.9911317 0.012851288348362968 1 +55 1 4.0095863 0.9821823 0.025937235909606105 1 +57 1 0.8893347 0.7087529 0.49664542168068465 1 +58 1 0.3305645 0.5818967 0.78116497512687677 1 +59 1 1.3888779 0.800413 0.321183473216892 1 +61 0 -6.533784 0.0014513859 0.0020954282545771739 0 +62 1 4.6103764 0.9901499 0.014281119626437895 1 +65 1 2.3961554 0.91653365 0.12574024706518463 1 +67 1 2.0528002 0.88623023 0.17424655523866805 1 +75 0 -5.281498 0.005059084 0.0073172402005513993 0 +78 0 -4.5923367 0.010027591 0.014539777931699532 0 +80 0 -4.096772 0.016354343 0.023789394001283711 0 +81 0 -4.654711 0.009426952 0.013664727433940642 0 +83 0 -3.514595 0.028899798 0.042307928774102763 0 +84 1 5.9317627 0.9973532 0.0038235931828602401 1 +85 1 2.8983269 0.9477637 0.077400717363867957 1 +86 1 0.6317673 0.6528901 0.61508795989775833 1 +87 1 3.859229 0.9793511 0.030101927408386315 1 +89 0 -5.927332 0.002658498 0.0038405091499816844 0 +94 0 -5.867672 0.0028214694 0.0040762732268601002 0 +101 1 -2.1323113 0.10599577 3.2379214462486434 0 +103 1 0.7015362 0.66852826 0.58093955032088451 1 +107 1 3.5016365 0.9707343 0.042851628370448841 1 +110 0 -4.524825 0.010720436 0.015549820680370292 0 +114 0 -4.1864657 0.014972332 0.021763846438684648 0 +116 0 -0.7707672 0.31631318 0.54859247573037895 0 +118 0 -6.563585 0.0014088316 0.0020339473944971918 0 +119 0 -4.879624 0.0075425496 0.010922843976522486 0 +124 1 5.156742 0.99427253 0.008286746215124376 1 +126 1 5.7151136 0.99671507 0.0047469549129191217 1 +127 0 -5.2313614 0.005317812 0.0076924524742365942 0 +130 0 -3.6252122 0.025951993 0.037935215641917444 0 +134 0 -5.580675 0.0037558598 0.005428761531925271 0 +135 0 -3.2147522 0.03861433 0.056812793866399591 0 +136 0 -4.9443645 0.0070730555 0.010240520805639591 0 +139 0 ? ? ? 0 +140 0 -5.929989 0.0026514626 0.0038303321306625395 0 +142 1 3.3667612 0.9666494 0.048935350664887807 1 +143 0 -5.9245596 0.0026658587 0.0038511567526827968 0 +146 1 -0.8934498 0.29039842 1.783894496397034 0 +148 0 -2.467762 0.07814931 0.11739499720015685 0 +149 1 7.946413 0.9996462 0.00051053483006131134 1 +153 0 -4.6546535 0.009427487 0.013665506008433347 0 +155 1 1.8597221 0.86526453 0.20878682350729794 1 +157 0 -5.867672 0.0028214694 0.0040762732268601002 0 +158 0 ? ? ? 0 +159 1 9.591595 0.9999317 9.8549424786689596E-05 1 +160 1 7.3637524 0.9993666 0.00091411935601581808 1 +162 0 -5.2313614 0.005317812 0.0076924524742365942 0 +163 0 -5.1793003 0.0056004045 0.0081023852955671421 0 +165 0 -4.0633006 0.016901605 0.024592276370398063 0 +166 1 5.4291134 0.9956322 0.0063152463458241801 1 +168 0 -5.2313614 0.005317812 0.0076924524742365942 0 +170 0 -5.929989 0.0026514626 0.0038303321306625395 0 +172 0 -6.8532963 0.0010548553 0.0015226377274540548 0 +175 1 4.9009466 0.9926154 0.010693254751102175 1 +178 0 -4.595051 0.010000683 0.014500565544495215 0 +182 0 -3.3224301 0.034809668 0.051114629699500667 0 +184 1 4.293844 0.98653156 0.019562896492565776 1 +185 0 -5.867672 0.0028214694 0.0040762732268601002 0 +186 1 3.0071125 0.9528944 0.069611767426840221 1 +190 1 10.157183 0.9999612 5.5981438976225112E-05 1 +193 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +194 0 -5.2313614 0.005317812 0.0076924524742365942 0 +195 0 -4.595051 0.010000683 0.014500565544495215 0 +197 0 -3.0897799 0.0435308 0.064209582068205179 0 +200 1 8.110399 0.9996997 0.00043328935662039213 1 +203 0 -3.9587402 0.01872965 0.027277425706471844 0 +208 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +213 1 11.922829 0.9999934 9.5450686985595153E-06 1 +214 1 11.574503 0.9999906 1.3586693394834943E-05 1 +215 1 6.340664 0.99824 0.0025413889437704167 1 +217 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +220 0 -6.214329 0.0019965644 0.0028833129439020211 0 +221 1 7.8691406 0.99961776 0.00055156779361044005 1 +222 1 -3.087008 0.043646354 4.5179950389321171 0 +224 1 8.318869 0.9997562 0.00035174739777412068 1 +225 0 -6.8532963 0.0010548553 0.0015226377274540548 0 +227 1 6.0467033 0.99763995 0.0034088506365501682 1 +229 1 10.046564 0.99995667 6.2517048110806339E-05 1 +230 1 4.703559 0.9910184 0.01301622995065306 1 +231 1 5.9737797 0.99746186 0.0036664235381914348 1 +232 0 1.1070662 0.7515817 2.0091568050125161 1 +234 0 -3.7643013 0.022658495 0.033065335000127247 0 +235 0 ? ? ? 0 +236 1 8.480522 0.9997926 0.00029928085261473974 1 +238 1 9.448765 0.9999212 0.00011368501131740873 1 +243 0 -4.5273056 0.010694161 0.01551150300683355 0 +245 0 -3.349574 0.033909112 0.04976917374708114 0 +251 1 7.3142223 0.99933445 0.00096049880304389836 1 +253 1 5.392255 0.9954689 0.0065518292521607709 1 +255 1 2.1359434 0.8943479 0.16109193821531068 1 +256 0 -5.929989 0.0026514626 0.0038303321306625395 0 +261 1 7.9640884 0.9996524 0.00050158859464882689 1 +263 1 7.3738213 0.9993729 0.00090499852703081662 1 +264 1 3.3298283 0.96543807 0.050744380486665137 1 +265 0 -2.72224 0.061673712 0.091838410690412098 0 +266 1 5.6310873 0.99642813 0.0051623412667997489 1 +270 1 4.851344 0.99224275 0.011234973837165527 1 +273 1 -0.24661922 0.4386558 1.1888387714139719 0 +274 0 -5.0191236 0.006566908 0.0095052903656552609 0 +281 0 -5.578018 0.0037658145 0.0054431773063631141 0 +282 1 1.5178432 0.82022065 0.28591603058749171 1 +286 1 11.969519 0.9999937 9.115109290810302E-06 1 +287 0 -5.580675 0.0037558598 0.005428761531925271 0 +289 1 5.619444 0.99638647 0.0052226659316945189 1 +292 1 ? ? ? 0 +294 0 ? ? ? 0 +295 1 5.155838 0.9942674 0.0082941840883700522 1 +298 0 -2.7162113 0.062023513 0.092376336902790707 0 +302 1 12.148499 0.9999947 7.6532482629398447E-06 1 +305 1 6.674677 0.9987391 0.0018202082685338874 1 +306 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +307 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +310 0 -6.2169857 0.0019912773 0.0028756700051982267 0 +313 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +315 0 ? ? ? 0 +318 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +320 1 4.269272 0.9862011 0.020046221903610821 1 +322 0 -5.2313614 0.005317812 0.0076924524742365942 0 +324 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +325 0 -4.782553 0.008305038 0.012031668786146766 0 +326 1 3.4337645 0.96874326 0.045813720120018366 1 +330 1 4.942194 0.9929117 0.010262676007266741 1 +334 1 4.4339123 0.98827124 0.017021043225296722 1 +335 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +337 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +339 1 4.2527885 0.98597497 0.020377075821022474 1 +340 1 5.304158 0.9950537 0.0071536968970724989 1 +341 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +342 0 -6.566299 0.0014050183 0.0020284382140880837 0 +345 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +351 0 -5.867672 0.0028214694 0.0040762732268601002 0 +356 1 -1.783071 0.14392434 2.7966174923741964 0 +357 1 9.025335 0.9998797 0.00017354093187002464 1 +359 1 4.652809 0.9905553 0.013690595592932819 1 +362 0 -4.5680046 0.010272039 0.014896057971768207 0 +363 0 -2.8947487 0.052413765 0.077670853816081298 0 +364 0 -5.867672 0.0028214694 0.0040762732268601002 0 +365 0 -6.2169857 0.0019912773 0.0028756700051982267 0 +367 1 8.16086 0.9997145 0.00041195725980106571 1 +369 0 -6.170454 0.0020859265 0.0030124988027857122 0 +372 0 -4.6029434 0.0099228425 0.014387134935430685 0 +374 0 -5.239254 0.005276226 0.0076321372386148386 0 +375 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +380 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +382 0 -4.355824 0.012669291 0.018394695292285344 0 +385 0 -4.6573496 0.009402342 0.013628884977755662 0 +386 1 4.0342827 0.9826094 0.02531006728743632 1 +390 0 -6.5636425 0.0014087503 0.0020338299987495033 0 +393 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +394 0 -6.1382294 0.0021540925 0.0031110505707663445 0 +397 0 -5.2936783 0.0049981424 0.0072288757659781672 0 +400 1 5.540056 0.99608904 0.0056533831124207911 1 +401 0 -5.929989 0.0026514626 0.0038303321306625395 0 +402 0 -3.5607624 0.027631931 0.040425576532739718 0 +403 0 -4.9714518 0.006885339 0.0099677998445315361 0 +405 0 -6.8532963 0.0010548553 0.0015226377274540548 0 +407 0 -6.8532963 0.0010548553 0.0015226377274540548 0 +408 0 -4.852635 0.0077472883 0.011220494995702469 0 +410 0 -6.8532963 0.0010548553 0.0015226377274540548 0 +411 0 ? ? ? 0 +412 1 7.5879498 0.9994937 0.0007305952773892084 1 +417 0 -6.8532963 0.0010548553 0.0015226377274540548 0 +420 0 -3.6393838 0.025596153 0.037408265779232425 0 +421 1 9.157963 0.9998946 0.00015204067460350784 1 +424 0 -5.929989 0.0026514626 0.0038303321306625395 0 +425 1 11.787142 0.9999924 1.1006931643385188E-05 1 +426 0 -2.8858137 0.052859314 0.078349358498549668 0 +427 1 3.751421 0.97705454 0.033489003404877538 1 +431 0 -3.1671953 0.040419057 0.059523588495218799 0 +432 0 -4.095816 0.01636973 0.023811962465425877 0 +433 0 -4.9770355 0.0068472624 0.0099124870448244508 0 +435 1 6.095413 0.9977519 0.0032469859785609744 1 +437 0 -5.2936783 0.0049981424 0.0072288757659781672 0 +438 0 -4.0315294 0.017437696 0.02537920439874838 0 +443 0 -6.883097 0.0010239151 0.0014779540602254402 0 +444 0 -2.8597865 0.05417764 0.08035884338728036 0 +445 0 -6.566299 0.0014050183 0.0020284382140880837 0 +446 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +447 0 -4.6573677 0.009402174 0.013628639475260666 0 +448 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +458 0 -4.3159466 0.013177927 0.01913810852238803 0 +459 0 -3.974526 0.01844172 0.026854164571853627 0 +460 0 -4.07806 0.016658103 0.024234982276436657 0 +461 0 -4.6519394 0.009452866 0.013702469790715719 0 +462 0 -3.4417505 0.03101583 0.045454998095521774 0 +463 0 -5.08144 0.006172621 0.00893280780221151 0 +468 0 -5.2936783 0.0049981424 0.0072288757659781672 0 +469 0 -6.511875 0.0014834872 0.0021418086899371901 0 +470 0 -5.580675 0.0037558598 0.005428761531925271 0 +471 0 -3.4417505 0.03101583 0.045454998095521774 0 +472 0 -4.021057 0.017618034 0.025644018148754801 0 +473 0 -5.2936783 0.0049981424 0.0072288757659781672 0 +475 0 -5.929989 0.0026514626 0.0038303321306625395 0 +476 0 -4.952257 0.0070178406 0.01016029739708614 0 +477 0 -5.2936783 0.0049981424 0.0072288757659781672 0 +478 0 -4.6251426 0.009707106 0.01407280825431988 0 +479 1 5.407748 0.9955383 0.0064512832547389109 1 +481 0 -3.382536 0.03284574 0.048182077539649704 0 +485 0 -5.9842777 0.0025117071 0.0036281858070808279 0 +486 0 -5.580675 0.0037558598 0.005428761531925271 0 +488 1 -0.16335392 0.4592521 1.1226418103257341 0 +490 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +491 1 4.140028 0.98432714 0.022790225287978638 1 +494 0 -0.08170223 0.4795858 0.94226775753788861 0 +496 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +498 0 -4.9443645 0.0070730555 0.010240520805639591 0 +499 0 -4.9443645 0.0070730555 0.010240520805639591 0 +500 0 -3.3224301 0.034809668 0.051114629699500667 0 +503 0 -4.595051 0.010000683 0.014500565544495215 0 +505 0 -5.8655386 0.002827478 0.0040849664452188555 0 +506 1 8.1275425 0.9997048 0.00042597791694096366 1 +508 0 -4.6573677 0.009402174 0.013628639475260666 0 +509 0 -6.566299 0.0014050183 0.0020284382140880837 0 +511 0 -4.308054 0.013280959 0.019288745432599251 0 +512 0 -4.6573677 0.009402174 0.013628639475260666 0 +515 1 5.338791 0.99522126 0.0069107941620064261 1 +516 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +518 0 -5.298914 0.004972172 0.0071912206884759524 0 +524 0 -5.580675 0.0037558598 0.005428761531925271 0 +525 0 -5.640335 0.0035391084 0.0051149106532798576 0 +526 0 -5.2936783 0.0049981424 0.0072288757659781672 0 +530 1 4.63558 0.99039274 0.013927349249580143 1 +536 0 -3.9587402 0.01872965 0.027277425706471844 0 +537 0 -3.6173196 0.026152255 0.038231861309927909 0 +542 0 -4.5929174 0.010021827 0.014531378095896472 0 +543 0 -4.9443645 0.0070730555 0.010240520805639591 0 +545 0 -4.308054 0.013280959 0.019288745432599251 0 +550 0 -5.580675 0.0037558598 0.005428761531925271 0 +551 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +552 0 -4.0065107 0.017871574 0.026016406332133601 0 +553 0 -3.0457807 0.045399986 0.067031737652501561 0 +554 0 -5.929989 0.0026514626 0.0038303321306625395 0 +555 0 -2.6956224 0.06323216 0.094236548790655558 0 +556 0 -3.6447544 0.025462545 0.037210461021787392 0 +562 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +564 0 -4.382813 0.012336095 0.017907908223938711 0 +567 0 -4.30534 0.0133165745 0.019340820470559086 0 +568 1 2.6288166 0.9326933 0.10052533697601265 1 +570 1 4.9678555 0.99309003 0.010003576488368728 1 +571 1 8.770479 0.9998448 0.00022393879069825088 1 +572 0 -5.580675 0.0037558598 0.005428761531925271 0 +573 0 -6.8532963 0.0010548553 0.0015226377274540548 0 +574 1 5.0986385 0.993932 0.0087809292203439634 1 +575 0 -3.6173196 0.026152255 0.038231861309927909 0 +576 0 -4.308054 0.013280959 0.019288745432599251 0 +579 0 -6.5039825 0.0014952245 0.0021587673525068959 0 +580 0 -3.9666328 0.018585142 0.027064981455609802 0 +583 0 -5.929989 0.0026514626 0.0038303321306625395 0 +585 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +587 0 -4.095816 0.01636973 0.023811962465425877 0 +588 1 4.5075893 0.9890952 0.015818692337728486 1 +589 0 -4.6573677 0.009402174 0.013628639475260666 0 +591 1 3.0920753 0.95656466 0.064065595221762014 1 +592 1 4.6950636 0.9909425 0.013126780011512462 1 +595 0 -4.308054 0.013280959 0.019288745432599251 0 +596 0 -4.6029434 0.0099228425 0.014387134935430685 0 +597 0 -3.275899 0.036407314 0.053504652181848217 0 +598 0 -5.580675 0.0037558598 0.005428761531925271 0 +599 0 -3.6008801 0.026574217 0.038857106970548651 0 +601 0 -6.8532963 0.0010548553 0.0015226377274540548 0 +603 1 3.5293217 0.97151065 0.041698285610926963 1 +605 1 8.48819 0.9997942 0.00029695860700686236 1 +608 1 7.3317604 0.999346 0.0009438054722544462 1 +610 1 7.064516 0.9991458 0.001232868314212441 1 +611 1 4.9259853 0.9927967 0.010429747129740782 1 +615 0 -4.261523 0.013904746 0.020201080857870868 0 +616 0 -5.580675 0.0037558598 0.005428761531925271 0 +620 0 -5.580675 0.0037558598 0.005428761531925271 0 +623 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +625 0 -4.28861 0.013538191 0.019664897126474463 0 +626 1 3.0649128 0.955422 0.06579002138022734 1 +628 0 -6.566299 0.0014050183 0.0020284382140880837 0 +630 0 -3.3059907 0.035366245 0.051946799741171994 0 +631 0 -4.308054 0.013280959 0.019288745432599251 0 +632 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +635 0 -4.9057264 0.0073496453 0.010642453973094183 0 +636 1 8.309149 0.99975383 0.00035518789362715982 1 +637 0 -2.9890575 0.04792267 0.070849338724798414 0 +640 0 -4.7143707 0.00888584 0.012876853170267714 0 +643 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +646 0 -6.8289638 0.0010808093 0.0015601214335835298 0 +647 0 -6.6233025 0.0013272706 0.0019161186585213146 0 +648 1 8.327747 0.99975836 0.0003486509585214664 1 +650 0 -4.681798 0.009177342 0.013301234651982449 0 +651 0 -6.241764 0.0019426377 0.0028053595949573376 0 +655 0 -5.580675 0.0037558598 0.005428761531925271 0 +658 1 6.5173016 0.99852455 0.0021302011627110694 1 +659 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +662 0 -6.170454 0.0020859265 0.0030124988027857122 0 +663 0 -6.170454 0.0020859265 0.0030124988027857122 0 +664 0 -5.001368 0.006683762 0.0096749993949486468 0 +666 0 -3.8116298 0.021633744 0.031553449287558002 0 +667 0 -5.2313614 0.005317812 0.0076924524742365942 0 +669 1 6.0170918 0.9975692 0.0035111674317070724 1 +671 0 -4.909483 0.00732229 0.010602697001513207 0 +672 0 -5.867672 0.0028214694 0.0040762732268601002 0 +673 0 -4.2514963 0.014042894 0.020403211794038669 0 +674 0 -6.8532963 0.0010548553 0.0015226377274540548 0 +675 0 -4.269416 0.013796935 0.020043358111662083 0 +676 0 -6.511875 0.0014834872 0.0021418086899371901 0 +677 0 -4.6573677 0.009402174 0.013628639475260666 0 +684 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +686 0 -7.2026095 0.00074408605 0.0010738888366935164 0 +687 0 -5.1561995 0.0057305344 0.0082911928533043935 0 +690 0 -6.6233025 0.0013272706 0.0019161186585213146 0 +695 0 -6.566299 0.0014050183 0.0020284382140880837 0 diff --git a/test/BaselineOutput/Common/LogisticRegression/net9.0/LogisticRegression-non-negative-TrainTest-breast-cancer.txt b/test/BaselineOutput/Common/LogisticRegression/net9.0/LogisticRegression-non-negative-TrainTest-breast-cancer.txt new file mode 100644 index 0000000000..78f075c546 --- /dev/null +++ b/test/BaselineOutput/Common/LogisticRegression/net9.0/LogisticRegression-non-negative-TrainTest-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +0 0 -3.9978466 0.018024284 0.026240747950219304 0 +1 0 2.2570772 0.90525925 3.3998711212270036 1 +2 0 -4.6860504 0.009138755 0.013245050133525319 0 +3 0 1.2745504 0.7815207 2.1944316584470722 1 +4 0 -3.8983502 0.019872414 0.028958533635626251 0 +5 1 10.383725 0.99996907 4.4630188215668621E-05 1 +6 0 -2.6780648 0.064280175 0.095851474393631592 0 +7 0 -5.310399 0.004915672 0.0071093033086943227 0 +8 0 -4.983145 0.006805839 0.009852314993309505 0 +9 0 -4.9553566 0.0069962745 0.010128964436990571 0 +10 0 -6.2179294 0.0019894028 0.0028729602518634948 0 +11 0 -6.039336 0.0023774744 0.0034340544033282191 0 +12 1 -0.8686323 0.29553896 1.7585797566458723 0 +13 0 -5.3742547 0.004612993 0.0066705395329983343 0 +14 1 8.055356 0.9996827 0.00045780445418641197 1 +15 1 0.7239523 0.67347676 0.57029994152617691 1 +16 0 -4.965962 0.00692298 0.010022481569704231 0 +17 0 -4.5345335 0.010617961 0.015400385742932211 0 +18 1 6.5412645 0.9985594 0.0020798227846953346 1 +19 0 -3.4611592 0.030437805 0.044594647933945998 0 +20 1 6.597213 0.9986377 0.0019667577188335239 1 +21 1 6.909402 0.99900264 0.0014396108762453246 1 +22 0 -5.502649 0.0040594144 0.0058684162356788536 0 +23 1 ? ? ? 0 +24 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +25 1 0.99812794 0.73069036 0.45266792076180506 1 +26 0 -5.565378 0.0038135338 0.005512283650499876 0 +27 0 -4.4292746 0.01178265 0.017099709536043457 0 +28 0 -6.039336 0.0023774744 0.0034340544033282191 0 +29 0 -6.412434 0.0016383367 0.0023655585370621285 0 +30 0 -5.5759835 0.0037734555 0.0054542427037081462 0 +31 0 -5.607908 0.0036553268 0.0052831836009057101 0 +32 1 6.295703 0.99815917 0.0026582034961301652 1 +33 0 -5.289816 0.005017387 0.007256779952781873 0 +34 0 -5.20514 0.0054583247 0.007896267805316699 0 +35 0 -6.039336 0.0023774744 0.0034340544033282191 0 +36 1 7.935828 0.99964243 0.00051595421118705818 1 +37 0 -1.7981739 0.1420735 0.22107403522003768 0 +38 1 4.9256725 0.99279445 0.010433038512589927 1 +39 1 0.8958311 0.71009207 0.49392200403100106 1 +40 0 ? ? ? 0 +41 1 2.722333 0.93833166 0.091830145633206933 1 +42 1 6.4560423 0.99843144 0.0022647243580549464 1 +43 1 -0.57765007 0.3594735 1.4760426875728427 0 +44 1 7.035142 0.99912035 0.0012696184696093834 1 +45 0 -6.2041183 0.0020170135 0.0029128740094185707 0 +46 1 2.9662752 0.9510271 0.072441649990985069 1 +47 0 -6.576023 0.0013914409 0.0020088227357911541 0 +48 0 -3.8983502 0.019872414 0.028958533635626251 0 +49 1 4.0427523 0.9827536 0.025098388304791521 1 +50 1 2.1004152 0.8909435 0.16659410599007651 1 +51 1 -0.5967989 0.3550764 1.4937986095905629 0 +52 1 4.9548855 0.99300045 0.010133726613379825 1 +53 1 7.151284 0.99921674 0.0011304547879411168 1 +54 1 6.3704157 0.9982915 0.002466963365520171 1 +55 1 3.6747484 0.9752712 0.036124603704880104 1 +56 1 4.184516 0.9849989 0.021806008082998943 1 +57 1 1.388957 0.8004257 0.32116059002187763 1 +58 1 1.2305021 0.77390647 0.36976887524319574 1 +59 1 0.9474764 0.7206074 0.47271462796496655 1 +60 1 1.7859716 0.8564327 0.22358825355591713 1 +61 0 -6.190853 0.0020438926 0.0029517313001556958 0 +62 1 4.6548157 0.990574 0.013663337123391597 1 +63 1 -0.14215755 0.46452034 1.1061863403755479 0 +64 0 -6.576023 0.0013914409 0.0020088227357911541 0 +65 1 3.702015 0.9759204 0.035164645125885144 1 +66 0 -4.5345335 0.010617961 0.015400385742932211 0 +67 1 3.0539465 0.9549526 0.066498973130157568 1 +68 1 10.140767 0.99996054 5.6927380571090205E-05 1 +69 0 -5.9195833 0.0026791224 0.0038703434501845806 0 +70 0 -3.8342571 0.021159967 0.030854988932352358 0 +71 1 6.778224 0.998863 0.0016413037970021442 1 +72 0 -3.2595491 0.036985263 0.054370219451433374 0 +73 1 7.999955 0.99966466 0.00048386833058238708 1 +74 1 1.8288069 0.86161953 0.21487714008834813 1 +75 0 -4.8054757 0.008118359 0.011760118092424837 0 +76 0 -5.683497 0.0033901103 0.0048992044840924943 0 +77 0 -4.089973 0.016464083 0.023950356040830144 0 +78 0 -4.3008804 0.013375294 0.019426681254675317 0 +79 0 -5.9660015 0.0025579152 0.0036950193206617783 0 +80 0 -3.6474137 0.02539664 0.037112897808120902 0 +81 0 -4.64787 0.009491046 0.01375807798387328 0 +82 0 -3.9872413 0.018212955 0.02651796554299252 0 +83 0 -2.9603357 0.049250282 0.072862489473219827 0 +84 1 8.486508 0.9997938 0.0002974746612633519 1 +85 1 5.1310053 0.9941241 0.0085021141109656376 1 +86 1 1.414465 0.8044692 0.31389085823029994 1 +87 1 5.359517 0.99531883 0.0067693573723155084 1 +88 0 -4.5345335 0.010617961 0.015400385742932211 0 +89 0 -5.64791 0.0035124945 0.0050763789997044584 0 +90 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +91 0 -5.615986 0.0036260241 0.0052407543160890172 0 +92 0 -4.5345335 0.010617961 0.015400385742932211 0 +93 0 -6.576023 0.0013914409 0.0020088227357911541 0 +94 0 -5.607908 0.0036553268 0.0052831836009057101 0 +95 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +96 0 -6.1526732 0.0021232683 0.0030664853162433482 0 +97 0 -3.9978466 0.018024284 0.026240747950219304 0 +98 1 8.128423 0.9997051 0.0004255478334078568 1 +99 1 9.796372 0.9999443 8.0318133658572531E-05 1 +100 1 4.421199 0.98812294 0.017237544994650796 1 +101 1 -1.7305994 0.15051092 2.7320599125487877 0 +102 0 -4.237025 0.014244678 0.020698501188506829 0 +103 1 1.0072308 0.73247784 0.44914297382199764 1 +104 1 11.243674 0.9999869 1.8918215632667518E-05 1 +105 1 2.2436552 0.90410185 0.14544279105572627 1 +106 1 7.5832148 0.99949133 0.00073403667682701737 1 +107 1 4.973281 0.99312717 0.0099496321405606308 1 +108 0 -6.1260004 0.0021805386 0.00314928705339875 0 +109 1 5.6273623 0.99641484 0.0051815862006337095 1 +110 0 -3.7353377 0.023308842 0.034025658135015208 0 +111 1 2.4953146 0.9238127 0.11432773497553032 1 +112 1 8.550721 0.99980664 0.00027898283227574002 1 +113 1 7.5348063 0.9994661 0.00077042997831079868 1 +114 0 -3.9821615 0.018304009 0.026651771128502508 0 +115 0 -5.466772 0.0042070732 0.0060823270953819231 0 +116 0 -1.161375 0.23841752 0.39292780679091621 0 +117 1 8.403041 0.9997759 0.00032336362004626066 1 +118 0 -6.0904427 0.0022592903 0.0032631545246152418 0 +119 0 -4.6754456 0.009235285 0.013385605715716141 0 +120 0 -5.528994 0.003954282 0.0057161321379725601 0 +121 0 -4.1387587 0.01569245 0.022818934422793962 0 +122 1 7.8866014 0.9996244 0.00054201913815316223 1 +123 1 2.4328165 0.9192957 0.12139905819786539 1 +124 1 6.4366713 0.9984008 0.0023089940169017807 1 +125 0 -6.576023 0.0013914409 0.0020088227357911541 0 +126 1 7.348096 0.99935657 0.00092857512761946313 1 +127 0 -5.0712204 0.0062356307 0.0090242788802262375 0 +128 1 4.0983734 0.98367137 0.023751685975252283 1 +129 0 -6.420775 0.0016247506 0.0023459258577485818 0 +130 0 -3.8342571 0.021159967 0.030854988932352358 0 +131 0 -5.607908 0.0036553268 0.0052831836009057101 0 +132 1 6.818391 0.99890774 0.0015766522486064089 1 +133 0 -5.2776375 0.0050785528 0.0073454708852408656 0 +134 0 -5.4293146 0.004366947 0.0063139690362318162 0 +135 0 -3.162654 0.040595565 0.059788986231469908 0 +136 0 -4.965962 0.00692298 0.010022481569704231 0 +137 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +138 0 -4.668453 0.009299485 0.013479092815885742 0 +139 0 ? ? ? 0 +140 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +141 0 -6.4707646 0.0015456489 0.0022316251099291748 0 +142 1 3.648243 0.97462386 0.037082554548014866 1 +143 0 -5.466772 0.0042070732 0.0060823270953819231 0 +144 0 -6.039336 0.0023774744 0.0034340544033282191 0 +145 0 ? ? ? 0 +146 1 -0.110322 0.47244743 1.0817743025455977 0 +147 0 -6.1175184 0.0021990717 0.0031760833238582602 0 +148 0 -1.5570927 0.17406422 0.2758984827724662 0 +149 1 10.011209 0.9999551 6.475292119065462E-05 1 +150 0 -6.2179294 0.0019894028 0.0028729602518634948 0 +151 1 4.0202293 0.98236763 0.025665064636613621 1 +152 1 8.323649 0.99975735 0.00035011316511816354 1 +153 0 -4.6270504 0.009688783 0.014046115114840997 0 +154 0 -7.007452 0.0009042934 0.001305209812438018 0 +155 1 2.364932 0.9141138 0.12955428327160223 1 +156 0 -6.2679095 0.0018925996 0.0027330311789211504 0 +157 0 -5.607908 0.0036553268 0.0052831836009057101 0 +158 0 ? ? ? 0 +159 1 10.917714 0.9999819 2.614159977229895E-05 1 +160 1 7.7599115 0.9995737 0.00061514108379339587 1 +161 0 -4.309522 0.013261733 0.019260634789626564 0 +162 0 -5.0712204 0.0062356307 0.0090242788802262375 0 +163 0 -3.7609248 0.022733388 0.033175892302694909 0 +164 0 ? ? ? 0 +165 0 -3.9243522 0.019372234 0.028222483652424368 0 +166 1 6.742324 0.9988215 0.0017012231317571679 1 +167 1 7.940713 0.99964416 0.0005134595729021004 1 +168 0 -5.0712204 0.0062356307 0.0090242788802262375 0 +169 0 -6.762695 0.0011547745 0.0016669500343684006 0 +170 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +171 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +172 0 -6.576023 0.0013914409 0.0020088227357911541 0 +173 1 13.357762 0.9999984 2.3217676548882582E-06 1 +174 1 4.6437006 0.9904697 0.013815261909217594 1 +175 1 6.036124 0.9976149 0.0034450528853451077 1 +176 0 -5.607908 0.0036553268 0.0052831836009057101 0 +177 1 3.483139 0.9702042 0.043639707825973283 1 +178 0 -4.5345335 0.010617961 0.015400385742932211 0 +179 1 1.0144339 0.73388696 0.44637023739379939 1 +180 0 -6.2179294 0.0019894028 0.0028729602518634948 0 +181 0 -7.007452 0.0009042934 0.001305209812438018 0 +182 0 -3.4611592 0.030437805 0.044594647933945998 0 +183 1 7.37292 0.99937236 0.00090577293479888533 1 +184 1 4.5937204 0.9899861 0.014519794301737114 1 +185 0 -5.6812425 0.0033977358 0.0049102431276932578 0 +186 1 4.237856 0.985767 0.020681399347574153 1 +187 1 13.261174 0.9999983 2.4937505927785606E-06 1 +188 1 7.3616076 0.9993652 0.00091609841142772008 1 +189 0 -5.538972 0.0039151767 0.0056594922477700288 0 +190 1 10.425599 0.9999703 4.2824315647486083E-05 1 +191 1 9.9119005 0.9999504 7.1546556808546518E-05 1 +192 0 -4.4292746 0.01178265 0.017099709536043457 0 +193 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +194 0 -5.0712204 0.0062356307 0.0090242788802262375 0 +195 0 -4.5345335 0.010617961 0.015400385742932211 0 +196 0 5.939311 0.9973731 8.5724252524349094 1 +197 0 -3.0435715 0.045495823 0.067176583914075813 0 +198 0 -7.007452 0.0009042934 0.001305209812438018 0 +199 0 -5.502649 0.0040594144 0.0058684162356788536 0 +200 1 9.00109 0.99987674 0.00017784102177338968 1 +201 1 8.007012 0.999667 0.0004805135478112585 1 +202 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +203 0 -3.9978466 0.018024284 0.026240747950219304 0 +204 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +205 1 10.318215 0.999967 4.7639980852592516E-05 1 +206 1 4.6046877 0.9900943 0.014362149941655099 1 +207 0 -6.2179294 0.0019894028 0.0028729602518634948 0 +208 0 -6.2179294 0.0019894028 0.0028729602518634948 0 +209 0 -4.071181 0.016771166 0.02440086982171049 0 +210 1 12.976369 0.9999977 3.3536655897601792E-06 1 +211 1 7.935506 0.9996423 0.00051612625536571392 1 +212 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +213 1 13.021174 0.9999978 3.1816825493598248E-06 1 +214 1 12.006362 0.9999939 8.7711418568707846E-06 1 +215 1 6.687505 0.99875516 0.0017970475849021653 1 +216 0 -6.576023 0.0013914409 0.0020088227357911541 0 +217 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +218 1 6.3308897 0.9982227 0.0025663706119391895 1 +219 0 -2.8661413 0.053852923 0.079863629656436286 0 +220 0 -5.8265033 0.0029397013 0.0042473385227374417 0 +221 1 8.934205 0.9998682 0.00019013934964380927 1 +222 1 -2.8624334 0.054042164 4.2097707380854708 0 +223 1 4.411948 0.98801386 0.017396809399230939 1 +224 1 8.94608 0.99986976 0.00018790328223314414 1 +225 0 -6.576023 0.0013914409 0.0020088227357911541 0 +226 1 9.03777 0.9998812 0.0001713908917246428 1 +227 1 6.0392904 0.99762243 0.0034341921153200391 1 +228 0 -6.2179294 0.0019894028 0.0028729602518634948 0 +229 1 10.761678 0.9999788 3.0613236649743679E-05 1 +230 1 5.0817976 0.99382955 0.0089296584157742172 1 +231 1 6.9891634 0.99907905 0.001329264157056246 1 +232 0 0.7225437 0.67316693 1.6133741305691605 1 +233 1 5.2977686 0.9950222 0.007199413154900987 1 +234 0 -3.5541105 0.027811218 0.040691607873554171 0 +235 0 ? ? ? 0 +236 1 10.420669 0.9999702 4.2996303413732479E-05 1 +237 1 5.82882 0.9970671 0.004237506305358023 1 +238 1 11.358827 0.9999883 1.6854398235588073E-05 1 +239 1 4.8551636 0.99227214 0.011192249316821019 1 +240 0 -2.3628817 0.0860473 0.12980859054058355 0 +241 0 -4.508372 0.010896344 0.015806374166766415 0 +242 0 -5.607908 0.0036553268 0.0052831836009057101 0 +243 0 -3.7410574 0.023178985 0.033833857076164139 0 +244 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +245 0 -3.2945848 0.035757434 0.052531976141166316 0 +246 1 9.59921 0.9999322 9.7775450197580527E-05 1 +247 1 2.0679398 0.8877478 0.17177817558162831 1 +248 0 -3.4094768 0.0320006 0.046921943545302838 0 +249 0 ? ? ? 0 +250 0 -6.699338 0.0012302118 0.0017759130792554746 0 +251 1 7.9959764 0.9996633 0.00048584679587356108 1 +252 0 3.6111555 0.9736903 5.2482598221939263 1 +253 1 6.4560423 0.99843144 0.0022647243580549464 1 +254 1 4.6548157 0.990574 0.013663337123391597 1 +255 1 2.5990705 0.93080175 0.10345417292552063 1 +256 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +257 0 -5.502649 0.0040594144 0.0058684162356788536 0 +258 0 -5.0712204 0.0062356307 0.0090242788802262375 0 +259 0 1.4420652 0.80877423 2.3866511593610102 1 +260 1 8.819928 0.99985224 0.00021318824647088272 1 +261 1 10.05152 0.9999569 6.2173067944539402E-05 1 +262 1 8.008122 0.99966735 0.0004799974280773332 1 +263 1 7.9955263 0.9996632 0.00048601883646188653 1 +264 1 4.3294287 0.98699623 0.018883515751996462 1 +265 0 -2.8803086 0.053135607 0.078770273333323801 0 +266 1 5.6979837 0.9966585 0.0048288319569100256 1 +267 1 1.7236948 0.84860414 0.23683637337760491 1 +268 1 8.499869 0.99979657 0.00029351825001431117 1 +269 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +270 1 5.1409435 0.9941819 0.0084182984458152921 1 +271 0 -3.9978466 0.018024284 0.026240747950219304 0 +272 1 1.7236948 0.84860414 0.23683637337760491 1 +273 1 -0.34247303 0.41520888 1.2680908096305508 0 +274 0 -4.846209 0.0077968426 0.011292546696238616 0 +275 0 ? ? ? 0 +276 0 -5.502649 0.0040594144 0.0058684162356788536 0 +277 0 -6.576023 0.0013914409 0.0020088227357911541 0 +278 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +279 1 6.586959 0.99862367 0.0019869933895667402 1 +280 0 -5.0712204 0.0062356307 0.0090242788802262375 0 +281 0 -5.289816 0.005017387 0.007256779952781873 0 +282 1 2.9461765 0.9500825 0.073875329200884868 1 +283 1 4.984359 0.9932024 0.0098403642172744926 1 +284 1 6.573718 0.9986054 0.0020134293529778819 1 +285 1 12.268481 0.9999953 6.7933307031989985E-06 1 +286 1 14.178321 0.9999993 1.031896274211761E-06 1 +287 0 -5.4293146 0.004366947 0.0063139690362318162 0 +288 1 1.7091665 0.84672815 0.2400292484255111 1 +289 1 7.1014795 0.9991768 0.0011881152907599552 1 +290 0 -7.007452 0.0009042934 0.001305209812438018 0 +291 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +292 1 ? ? ? 0 +293 1 4.6855135 0.9908564 0.013252091893358501 1 +294 0 ? ? ? 0 +295 1 6.2168865 0.99800855 0.0028759207527209812 1 +296 0 1.0209198 0.7351517 1.9167618908129638 1 +297 0 ? ? ? 0 +298 0 -3.020033 0.046529017 0.068739061834349816 0 +299 1 7.0977564 0.9991737 0.0011925905306341051 1 +300 1 6.6469717 0.9987037 0.0018713525078078205 1 +301 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +302 1 14.444997 0.99999946 7.7392213646475052E-07 1 +303 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +304 1 5.0605726 0.99369806 0.0091205456903274886 1 +305 1 7.3606424 0.9993646 0.00091695887114923895 1 +306 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +307 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +308 1 6.290221 0.9981491 0.0026727629048521676 1 +309 0 -2.6676831 0.06490745 0.096818927895736318 0 +310 0 -5.9660015 0.0025579152 0.0036950193206617783 0 +311 0 -7.007452 0.0009042934 0.001305209812438018 0 +312 1 2.9073153 0.9482069 0.076726201088877163 1 +313 0 -7.007452 0.0009042934 0.001305209812438018 0 +314 0 -6.649358 0.0012931789 0.0018668701168435479 0 +315 0 ? ? ? 0 +316 1 2.1206856 0.8928975 0.16343354545162167 1 +317 1 7.571534 0.9994854 0.00074264021133653045 1 +318 0 -5.924591 0.0026657751 0.0038510358409564985 0 +319 0 1.3538389 0.79475653 2.2845917848128998 1 +320 1 5.98847 0.99749875 0.0036130604489300233 1 +321 0 ? ? ? 0 +322 0 -5.0712204 0.0062356307 0.0090242788802262375 0 +323 1 4.282296 0.98637724 0.019788585123052597 1 +324 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +325 0 -4.7445126 0.008624276 0.012496163398747154 0 +326 1 3.690011 0.97563666 0.035584124242818566 1 +327 0 -6.576023 0.0013914409 0.0020088227357911541 0 +328 1 2.0947208 0.890389 0.16749228451966611 1 +329 1 6.571471 0.9986022 0.0020179932654013581 1 +330 1 4.8110256 0.9919262 0.011695317633988674 1 +331 0 -3.6259413 0.02593357 0.037907928352497858 0 +332 0 -3.5780582 0.027170995 0.039741851730161766 0 +333 1 3.6066217 0.9735739 0.038637568477387217 1 +334 1 4.67328 0.9907449 0.013414475487097705 1 +335 0 -7.007452 0.0009042934 0.001305209812438018 0 +336 1 4.286172 0.9864292 0.019712567087271423 1 +337 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +338 0 -6.649358 0.0012931789 0.0018668701168435479 0 +339 1 4.4650183 0.98862636 0.01650271964280297 1 +340 1 5.6191187 0.9963853 0.0052243919964213787 1 +341 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +342 0 -6.4707646 0.0015456489 0.0022316251099291748 0 +343 0 -7.007452 0.0009042934 0.001305209812438018 0 +344 1 8.533357 0.99980325 0.00028388529408090524 1 +345 0 -7.007452 0.0009042934 0.001305209812438018 0 +346 0 -3.7111406 0.023866102 0.034849036661125946 0 +347 0 -6.630764 0.001317417 0.0019018840585968357 0 +348 1 -0.61402607 0.35114133 1.5098762660253275 0 +349 1 2.9637518 0.95090944 0.072620148982770999 1 +350 0 -4.2245197 0.014421342 0.020957078957492121 0 +351 0 -5.607908 0.0036553268 0.0052831836009057101 0 +352 0 -0.5196886 0.37292504 0.67329019017635949 0 +353 1 7.176564 0.9992363 0.0011022277999205644 1 +354 0 -6.576023 0.0013914409 0.0020088227357911541 0 +355 0 -4.667616 0.009307203 0.013490331910073598 0 +356 1 -1.3441114 0.20683475 2.2734495150623477 0 +357 1 11.126069 0.9999853 2.1240013733369503E-05 1 +358 1 4.3297377 0.9870002 0.018877678437776931 1 +359 1 4.87932 0.9924552 0.01092613783835146 1 +360 1 13.757287 0.9999989 1.5478446880940214E-06 1 +361 1 4.2205095 0.98552155 0.021040669204676572 1 +362 0 -4.277526 0.013687018 0.019882572534073818 0 +363 0 -2.953189 0.049586013 0.073372026756355624 0 +364 0 -5.607908 0.0036553268 0.0052831836009057101 0 +365 0 -6.039336 0.0023774744 0.0034340544033282191 0 +366 1 12.063476 0.9999942 8.3411826797711399E-06 1 +367 1 9.634366 0.99993455 9.4421565109537245E-05 1 +368 0 -6.412434 0.0016383367 0.0023655585370621285 0 +369 0 -6.0543404 0.0023421517 0.0033829739255167063 0 +370 0 -4.5739083 0.010212191 0.014808822487048909 0 +371 0 -6.412434 0.0016383367 0.0023655585370621285 0 +372 0 -4.668453 0.009299485 0.013479092815885742 0 +373 0 -4.111183 0.016124126 0.023451777940682497 0 +374 0 -5.20514 0.0054583247 0.007896267805316699 0 +375 0 -7.007452 0.0009042934 0.001305209812438018 0 +376 0 -6.576023 0.0013914409 0.0020088227357911541 0 +377 0 -6.649358 0.0012931789 0.0018668701168435479 0 +378 0 -4.2500277 0.014063244 0.020432988220160278 0 +379 0 -2.9374404 0.05033348 0.074507103817010437 0 +380 0 -7.007452 0.0009042934 0.001305209812438018 0 +381 1 8.953799 0.9998708 0.00018644124003079258 1 +382 0 -4.220003 0.01448568 0.021051259775101649 0 +383 0 -6.4707646 0.0015456489 0.0022316251099291748 0 +384 0 -6.4707646 0.0015456489 0.0022316251099291748 0 +385 0 -4.255111 0.013992932 0.020330106120116161 0 +386 1 4.444525 0.9883936 0.016842418920701662 1 +387 0 -2.9901924 0.04787092 0.070770920788664884 0 +388 0 -5.814325 0.0029756138 0.0042993029971469827 0 +389 0 -3.745759 0.02307277 0.033676992912648515 0 +390 0 -6.2579317 0.0019115419 0.002760411200110962 0 +391 1 8.564824 0.9998093 0.0002751124794602528 1 +392 0 -5.502649 0.0040594144 0.0058684162356788536 0 +393 0 -7.0807867 0.0008404043 0.0012129568798654386 0 +394 0 -5.970401 0.0025467156 0.0036788203400675078 0 +395 0 -5.502649 0.0040594144 0.0058684162356788536 0 +396 0 -5.0712204 0.0062356307 0.0090242788802262375 0 +397 0 -5.3973904 0.004507969 0.0065183279877715399 0 +398 0 -5.2670326 0.0051324186 0.0074235816517891585 0 +399 0 -6.4018292 0.0016557745 0.0023907575424690201 0 +400 1 8.031964 0.9996752 0.00046864284064470924 1 +401 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +402 0 -3.8130474 0.02160376 0.031509234405850892 0 +403 0 -4.7785673 0.008337931 0.012079520761115079 0 +404 0 -6.135652 0.0021596393 0.0031190700996340543 0 +405 0 -6.576023 0.0013914409 0.0020088227357911541 0 +406 0 -4.6578484 0.009397699 0.013622122145295991 0 +407 0 -6.576023 0.0013914409 0.0020088227357911541 0 +408 0 -4.585927 0.01009142 0.014632799644806108 0 +409 0 -5.20514 0.0054583247 0.007896267805316699 0 +410 0 -6.576023 0.0013914409 0.0020088227357911541 0 +411 0 ? ? ? 0 +412 1 7.7017794 0.9995482 0.00065196153701951257 1 +413 0 -3.7003374 0.02411908 0.035222977147123087 0 +414 1 5.4194784 0.9955901 0.0063762238438268091 1 +415 0 -1.4183235 0.19492453 0.31280406913354142 0 +416 1 7.034707 0.99912 0.0012701348719051887 1 +417 0 -6.576023 0.0013914409 0.0020088227357911541 0 +418 0 -2.3895545 0.08397269 0.12653748802521594 0 +419 0 -6.229551 0.001966462 0.0028397980761373544 0 +420 0 -3.3565836 0.033680238 0.049427428639960608 0 +421 1 9.8025255 0.9999447 7.9802157073450512E-05 1 +422 0 -3.4682827 0.030228283 0.044282916538364352 0 +423 0 -3.8342571 0.021159967 0.030854988932352358 0 +424 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +425 1 12.883061 0.99999744 3.6976317320669667E-06 1 +426 0 -3.3519497 0.033831377 0.049653093260845861 0 +427 1 3.496297 0.97058225 0.043077623324427869 1 +428 0 -6.576023 0.0013914409 0.0020088227357911541 0 +429 0 -6.039336 0.0023774744 0.0034340544033282191 0 +430 0 -6.3298016 0.0017792162 0.0025691526422317577 0 +431 0 -2.9899826 0.047880482 0.070785410738434598 0 +432 0 -4.2042637 0.014712099 0.021382753388115749 0 +433 0 -4.9897804 0.006761135 0.0097873803440660882 0 +434 0 3.2537346 0.96280706 4.7488273686100584 1 +435 1 5.840127 0.99709994 0.0041899864942212704 1 +436 1 2.8956041 0.9476287 0.077606237188381638 1 +437 0 -5.3973904 0.004507969 0.0065183279877715399 0 +438 0 -4.1661897 0.015274326 0.022206222439164654 0 +439 0 -4.8607035 0.007685509 0.01113067322992304 0 +440 1 8.808446 0.9998506 0.00021559636141530833 1 +441 0 -2.4428043 0.07996635 0.12024146872291093 0 +442 0 -5.6009283 0.003680834 0.0053201183291922728 0 +443 0 -6.622282 0.0013286239 0.0019180735262927767 0 +444 0 -2.765482 0.05921822 0.088067975565166157 0 +445 0 -6.4707646 0.0015456489 0.0022316251099291748 0 +446 0 -7.007452 0.0009042934 0.001305209812438018 0 +447 0 -4.8607035 0.007685509 0.01113067322992304 0 +448 0 -7.0807867 0.0008404043 0.0012129568798654386 0 +449 1 8.61461 0.9998186 0.00026169533675417532 1 +450 0 -4.4031134 0.01209119 0.01755021591985494 0 +451 0 -4.8607035 0.007685509 0.01113067322992304 0 +452 0 -5.297894 0.0049772207 0.0071985408379837143 0 +453 1 7.0545015 0.9991372 0.0012452617046230653 1 +454 0 -6.3762197 0.0016986527 0.0024527216472044629 0 +455 1 0.42860126 0.6055396 0.72370673761693649 1 +456 1 9.124024 0.999891 0.00015728670782410122 1 +457 1 7.7657614 0.99957615 0.00061161394016591582 1 +458 0 -4.5631948 0.010321054 0.014967506732671972 0 +459 0 -4.265686 0.013847777 0.020117735438530297 0 +460 0 -4.2245197 0.014421342 0.020957078957492121 0 +461 0 -4.3933973 0.0122078 0.017720517622574369 0 +462 0 -3.6878328 0.024415161 0.035660757349862515 0 +463 0 -5.1723795 0.005639079 0.008158496230404556 0 +464 0 -5.3973904 0.004507969 0.0065183279877715399 0 +465 1 8.705016 0.9998343 0.00023907569277659222 1 +466 1 8.096089 0.99969536 0.00043956862262736329 1 +467 1 6.1876526 0.99794954 0.0029612245600862024 1 +468 0 -5.3973904 0.004507969 0.0065183279877715399 0 +469 0 -6.2785144 0.0018726724 0.002704227970268877 0 +470 0 -5.5759835 0.0037734555 0.0054542427037081462 0 +471 0 -3.6878328 0.024415161 0.035660757349862515 0 +472 0 -4.397351 0.012160217 0.017651024309070347 0 +473 0 -5.3973904 0.004507969 0.0065183279877715399 0 +474 0 -4.8607035 0.007685509 0.01113067322992304 0 +475 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +476 0 -5.099882 0.0060605113 0.0087700721407504827 0 +477 0 -5.3973904 0.004507969 0.0065183279877715399 0 +478 0 -4.8500986 0.0077668102 0.011248879349031599 0 +479 1 6.4265413 0.9983846 0.0023324213120842659 1 +480 0 -4.934038 0.0071459482 0.010346435825261097 0 +481 0 -3.5877094 0.02691705 0.039365302274136975 0 +482 1 14.304355 0.9999994 8.5991351058826534E-07 1 +483 1 9.330981 0.99991137 0.00012787476790525535 1 +484 0 -4.5631948 0.010321054 0.014967506732671972 0 +485 0 -5.674263 0.003421452 0.0049445753287763955 0 +486 0 -5.5759835 0.0037734555 0.0054542427037081462 0 +487 1 11.726071 0.9999919 1.1694867659388439E-05 1 +488 1 0.2558651 0.56361955 0.82720642995515892 1 +489 1 -1.4821672 0.18510029 2.433620961211663 0 +490 0 -7.007452 0.0009042934 0.001305209812438018 0 +491 1 4.6763144 0.99077266 0.013374029761460475 1 +492 0 -5.039296 0.0064366083 0.0093160770625755895 0 +493 1 7.9641085 0.9996524 0.00050158859464882689 1 +494 0 0.4184084 0.6031023 1.3331609887997158 1 +495 0 -5.5759835 0.0037734555 0.0054542427037081462 0 +496 0 -7.0807867 0.0008404043 0.0012129568798654386 0 +497 0 -5.3867855 0.004555811 0.0065876636843487562 0 +498 0 -4.965962 0.00692298 0.010022481569704231 0 +499 0 -4.965962 0.00692298 0.010022481569704231 0 +500 0 -3.4611592 0.030437805 0.044594647933945998 0 +501 0 -4.965962 0.00692298 0.010022481569704231 0 +502 0 -4.64787 0.009491046 0.01375807798387328 0 +503 0 -4.5345335 0.010617961 0.015400385742932211 0 +504 0 -7.007452 0.0009042934 0.001305209812438018 0 +505 0 -5.9128675 0.0026971265 0.0038963878796869182 0 +506 1 9.393357 0.99991673 0.00012013488337412622 1 +507 0 -5.851942 0.0028660737 0.004140807145729836 0 +508 0 -4.8607035 0.007685509 0.01113067322992304 0 +509 0 -6.4707646 0.0015456489 0.0022316251099291748 0 +510 0 -7.007452 0.0009042934 0.001305209812438018 0 +511 0 -4.4292746 0.01178265 0.017099709536043457 0 +512 0 -4.8607035 0.007685509 0.01113067322992304 0 +513 0 -5.5759835 0.0037734555 0.0054542427037081462 0 +514 1 9.33376 0.9999116 0.0001275307721554692 1 +515 1 6.515527 0.9985219 0.0021339903768075922 1 +516 0 -7.0807867 0.0008404043 0.0012129568798654386 0 +517 0 -6.649358 0.0012931789 0.0018668701168435479 0 +518 0 -5.318477 0.004876315 0.0070522441439271628 0 +519 1 4.4496126 0.98845184 0.016757421355254449 1 +520 0 -6.9341173 0.0009730347 0.0014044757609679298 0 +521 0 -5.3240557 0.00484932 0.0070131078224030885 0 +522 1 4.6514673 0.9905427 0.013708912879307096 1 +523 1 6.394397 0.9983319 0.0024085626488484649 1 +524 0 -5.502649 0.0040594144 0.0058684162356788536 0 +525 0 -5.615986 0.0036260241 0.0052407543160890172 0 +526 0 -5.3973904 0.004507969 0.0065183279877715399 0 +527 0 -4.5345335 0.010617961 0.015400385742932211 0 +528 0 -3.4109073 0.03195632 0.046855946893667656 0 +529 0 -5.039296 0.0064366083 0.0093160770625755895 0 +530 1 5.343629 0.9952442 0.0068775289173476856 1 +531 0 -4.6578484 0.009397699 0.013622122145295991 0 +532 0 -6.2179294 0.0019894028 0.0028729602518634948 0 +533 0 -5.502649 0.0040594144 0.0058684162356788536 0 +534 0 -6.039336 0.0023774744 0.0034340544033282191 0 +535 0 -5.2314854 0.005317156 0.0076915015129353182 0 +536 0 -3.9978466 0.018024284 0.026240747950219304 0 +537 0 -3.7003374 0.02411908 0.035222977147123087 0 +538 0 -4.965962 0.00692298 0.010022481569704231 0 +539 0 -3.8925877 0.019984964 0.029124211179545535 0 +540 0 -4.0441046 0.01722354 0.025064794013241035 0 +541 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +542 0 -4.8394933 0.007848968 0.011368340380769089 0 +543 0 -4.965962 0.00692298 0.010022481569704231 0 +544 0 -5.126227 0.005903864 0.0085427174533614456 0 +545 0 -4.4292746 0.01178265 0.017099709536043457 0 +546 1 9.538707 0.999928 0.00010388126101663741 1 +547 0 -6.5440993 0.0014365127 0.0020739397032742131 0 +548 0 -6.007412 0.0024544087 0.0035453158359960717 0 +549 1 5.013379 0.9933955 0.0095598727339306962 1 +550 0 -5.502649 0.0040594144 0.0058684162356788536 0 +551 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +552 0 -3.7885747 0.02212714 0.032281192719350892 0 +553 0 -2.724053 0.06156888 0.091677236297325188 0 +554 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +555 0 -2.5068154 0.07538177 0.11307028864257092 0 +556 0 -3.8764038 0.02030441 0.029594548427245702 0 +557 0 -4.2245197 0.014421342 0.020957078957492121 0 +558 0 -6.039336 0.0023774744 0.0034340544033282191 0 +559 0 -4.4292746 0.01178265 0.017099709536043457 0 +560 0 -3.9978466 0.018024284 0.026240747950219304 0 +561 0 -3.9978466 0.018024284 0.026240747950219304 0 +562 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +563 0 -5.502649 0.0040594144 0.0058684162356788536 0 +564 0 -4.309522 0.013261733 0.019260634789626564 0 +565 1 10.337827 0.99996763 4.6694045347237877E-05 1 +566 0 -4.7737117 0.008378175 0.012138070390060625 0 +567 0 -4.195622 0.014837893 0.021566957001074475 0 +568 1 3.1369467 0.9583913 0.06131326984745112 1 +569 1 9.030463 0.9998803 0.00017268091542736876 1 +570 1 5.843155 0.9971087 0.0041773090595500755 1 +571 1 9.603944 0.9999325 9.7345464494158371E-05 1 +572 0 -5.502649 0.0040594144 0.0058684162356788536 0 +573 0 -6.576023 0.0013914409 0.0020088227357911541 0 +574 1 4.7992887 0.99183166 0.011832816514707677 1 +575 0 -3.7003374 0.02411908 0.035222977147123087 0 +576 0 -4.4292746 0.01178265 0.017099709536043457 0 +577 0 -6.576023 0.0013914409 0.0020088227357911541 0 +578 0 -6.576023 0.0013914409 0.0020088227357911541 0 +579 0 -6.1445947 0.0021404538 0.0030913316669355846 0 +580 0 -4.1317663 0.015800823 0.02297778412444049 0 +581 1 6.6461153 0.99870265 0.0018729023615442059 1 +582 1 6.029276 0.9975985 0.0034687572308686456 1 +583 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +584 0 -3.1969104 0.039282158 0.057815313985279886 0 +585 0 -7.007452 0.0009042934 0.001305209812438018 0 +586 1 12.053037 0.99999416 8.427174504939984E-06 1 +587 0 -4.2042637 0.014712099 0.021382753388115749 0 +588 1 4.34554 0.9872014 0.018583665180518474 1 +589 0 -4.8607035 0.007685509 0.01113067322992304 0 +590 1 2.082511 0.8891917 0.16943363395548017 1 +591 1 4.2739954 0.98626524 0.019952403658026578 1 +592 1 4.842986 0.9921782 0.011328833565094636 1 +593 0 -4.5631948 0.010321054 0.014967506732671972 0 +594 1 3.0023432 0.9526799 0.069936585816428518 1 +595 0 -4.4292746 0.01178265 0.017099709536043457 0 +596 0 -4.668453 0.009299485 0.013479092815885742 0 +597 0 -3.4028287 0.032207176 0.047229853075843209 0 +598 0 -5.502649 0.0040594144 0.0058684162356788536 0 +599 0 -4.0309067 0.01744837 0.025394875573271941 0 +600 0 -5.502649 0.0040594144 0.0058684162356788536 0 +601 0 -6.649358 0.0012931789 0.0018668701168435479 0 +602 0 -4.965962 0.00692298 0.010022481569704231 0 +603 1 4.0770693 0.98332566 0.02425880386683087 1 +604 1 4.763097 0.99153316 0.012267073043375619 1 +605 1 9.002568 0.9998769 0.00017758301601775289 1 +606 0 -5.079299 0.006185769 0.0089518942180366601 0 +607 0 -7.007452 0.0009042934 0.001305209812438018 0 +608 1 8.711211 0.9998353 0.00023761359871469259 1 +609 0 -4.8607035 0.007685509 0.01113067322992304 0 +610 1 7.8127317 0.9995956 0.00058356915380086715 1 +611 1 5.3003664 0.995035 0.0071808326485285241 1 +612 1 15.40018 0.9999998 2.5797399935888651E-07 1 +613 0 -6.018126 0.002428316 0.0035075799987226873 0 +614 0 -6.112671 0.0022097342 0.0031915000145038542 0 +615 0 -4.3709445 0.0124815395 0.018120376980472964 0 +616 0 -5.502649 0.0040594144 0.0058684162356788536 0 +617 0 ? ? ? 0 +618 0 -4.965962 0.00692298 0.010022481569704231 0 +619 0 -4.4292746 0.01178265 0.017099709536043457 0 +620 0 -5.502649 0.0040594144 0.0058684162356788536 0 +621 0 -0.4559574 0.38794526 0.70826741747430166 0 +622 0 -2.4222083 0.0814948 0.12264020706614719 0 +623 0 -7.007452 0.0009042934 0.001305209812438018 0 +624 0 -4.184518 0.015001087 0.021805963068453948 0 +625 0 -4.110215 0.016139489 0.023474305616941124 0 +626 1 4.390518 0.98775744 0.017771280556490125 1 +627 0 -4.930085 0.007174049 0.010387269070340583 0 +628 0 -6.4707646 0.0015456489 0.0022316251099291748 0 +629 0 -5.3973904 0.004507969 0.0065183279877715399 0 +630 0 -3.7183933 0.023697723 0.034600198697996562 0 +631 0 -4.4292746 0.01178265 0.017099709536043457 0 +632 0 -7.007452 0.0009042934 0.001305209812438018 0 +633 1 3.4009504 0.9677342 0.047317220463578967 1 +634 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +635 0 -5.041551 0.0064222035 0.0092951608992207606 0 +636 1 8.948316 0.99987006 0.0001874732696668834 1 +637 0 -3.0414772 0.045586858 0.067314186129754053 0 +638 0 -5.3973904 0.004507969 0.0065183279877715399 0 +639 0 -4.2245197 0.014421342 0.020957078957492121 0 +640 0 -4.761207 0.008482704 0.012290155793123585 0 +641 0 -5.502649 0.0040594144 0.0058684162356788536 0 +642 0 -5.502649 0.0040594144 0.0058684162356788536 0 +643 0 -7.007452 0.0009042934 0.001305209812438018 0 +644 0 -6.4707646 0.0015456489 0.0022316251099291748 0 +645 0 -5.502649 0.0040594144 0.0058684162356788536 0 +646 0 -6.699338 0.0012302118 0.0017759130792554746 0 +647 0 -6.3712683 0.00170707 0.0024648859488124028 0 +648 1 11.002097 0.9999833 2.4077772041939022E-05 1 +649 0 -5.502649 0.0040594144 0.0058684162356788536 0 +650 0 -4.387141 0.0122834705 0.01783104107091768 0 +651 0 -5.929235 0.002653457 0.0038332171333820149 0 +652 0 -4.2042637 0.014712099 0.021382753388115749 0 +653 0 -4.965962 0.00692298 0.010022481569704231 0 +654 0 -5.0712204 0.0062356307 0.0090242788802262375 0 +655 0 -5.502649 0.0040594144 0.0058684162356788536 0 +656 0 -4.4292746 0.01178265 0.017099709536043457 0 +657 0 -1.2109308 0.2295364 0.37620129421616127 0 +658 1 7.3752756 0.99937385 0.00090362180313618902 1 +659 0 -7.007452 0.0009042934 0.001305209812438018 0 +660 0 -6.576023 0.0013914409 0.0020088227357911541 0 +661 0 -4.5345335 0.010617961 0.015400385742932211 0 +662 0 -5.9810057 0.0025199184 0.0036400619981428553 0 +663 0 -5.9810057 0.0025199184 0.0036400619981428553 0 +664 0 -4.8664656 0.0076416894 0.011066966603912513 0 +665 0 -7.007452 0.0009042934 0.001305209812438018 0 +666 0 -3.8625927 0.02058097 0.03000186680137043 0 +667 0 -5.0712204 0.0062356307 0.0090242788802262375 0 +668 1 1.7462759 0.85148245 0.23195129859587327 1 +669 1 6.51332 0.99851865 0.00213872690842372 1 +670 1 4.3528633 0.9872936 0.018448918239184674 1 +671 0 -4.7425237 0.008641298 0.012520934483732088 0 +672 0 -5.607908 0.0036553268 0.0052831836009057101 0 +673 0 -4.46865 0.011332875 0.016443234285328498 0 +674 0 -6.576023 0.0013914409 0.0020088227357911541 0 +675 0 -4.504864 0.010934212 0.015861609786469777 0 +676 0 -6.2785144 0.0018726724 0.002704227970268877 0 +677 0 -4.8607035 0.007685509 0.01113067322992304 0 +678 0 -7.007452 0.0009042934 0.001305209812438018 0 +679 0 -6.4707646 0.0015456489 0.0022316251099291748 0 +680 1 15.284507 0.99999976 3.4396534272948301E-07 1 +681 1 7.9829264 0.9996589 0.00049221231130512611 1 +682 0 -3.7728348 0.022470288 0.032787541076677536 0 +683 0 -7.007452 0.0009042934 0.001305209812438018 0 +684 0 -7.007452 0.0009042934 0.001305209812438018 0 +685 0 -7.007452 0.0009042934 0.001305209812438018 0 +686 0 -7.007452 0.0009042934 0.001305209812438018 0 +687 0 -5.052626 0.006351919 0.009193109868750755 0 +688 0 -5.3973904 0.004507969 0.0065183279877715399 0 +689 0 -4.404117 0.012079205 0.017532714784305702 0 +690 0 -6.3712683 0.00170707 0.0024648859488124028 0 +691 1 2.8413572 0.94487023 0.081811888611410546 1 +692 0 -5.9340773 0.0026406732 0.0038147250597758971 0 +693 0 -5.130744 0.005877412 0.0085043293130297008 0 +694 0 -5.475573 0.0041703633 0.0060291430713179054 0 +695 0 -6.4707646 0.0015456489 0.0022316251099291748 0 +696 1 5.1025248 0.9939554 0.0087470152271597566 1 +697 1 3.355424 0.966282 0.049483792818586683 1 +698 1 4.2770147 0.9863061 0.019892680536708389 1 diff --git a/test/BaselineOutput/Common/MulticlassLogisticRegression/net9.0/MulticlassLogisticRegression-CV-iris.txt b/test/BaselineOutput/Common/MulticlassLogisticRegression/net9.0/MulticlassLogisticRegression-CV-iris.txt new file mode 100644 index 0000000000..52fa88ab7d --- /dev/null +++ b/test/BaselineOutput/Common/MulticlassLogisticRegression/net9.0/MulticlassLogisticRegression-CV-iris.txt @@ -0,0 +1,151 @@ +Instance Label Assigned Log-loss #1 Score #2 Score #3 Score #1 Class #2 Class #3 Class +5 0 0 0.0039186265649996796 0.99608904 0.003910693 2.0860485E-11 0 1 2 +6 0 0 0.0066385190914096566 0.99338347 0.0066169878 2.15196E-10 0 1 2 +8 0 0 0.025494106553378942 0.9748281 0.02517165 2.8383844E-09 0 1 2 +9 0 0 0.019190772567240676 0.9809922 0.01900781 2.7052444E-10 0 1 2 +10 0 0 0.0035562493991592786 0.99645007 0.0035501744 4.9260778E-12 0 1 2 +11 0 0 0.012516964292635404 0.98756105 0.012439178 2.8002256E-10 0 1 2 +18 0 0 0.0047769616990695207 0.99523443 0.0047658556 8.125339E-12 0 1 2 +20 0 0 0.014707546996037419 0.9854001 0.014599827 8.912887E-11 0 1 2 +21 0 0 0.003647833406238951 0.9963588 0.0036412787 3.5076754E-11 0 1 2 +25 0 0 0.032316695740782254 0.9681999 0.03180029 1.1000117E-09 0 1 2 +28 0 0 0.0057745087199158104 0.99424213 0.005757904 2.0231521E-11 0 1 2 +31 0 0 0.0071621105155360203 0.9928635 0.007136584 6.2415725E-11 0 1 2 +32 0 0 0.0014066451993830036 0.99859434 0.0014055306 7.3224635E-13 0 1 2 +35 0 0 0.0052270768592320212 0.99478656 0.0052132453 2.502141E-11 0 1 2 +37 0 0 0.019190772567240676 0.9809922 0.01900781 2.7052444E-10 0 1 2 +40 0 0 0.0033063652339419308 0.9966991 0.0033008496 1.8953988E-11 0 1 2 +41 0 0 0.07708743004327992 0.9258089 0.07419081 3.333749E-08 0 1 2 +44 0 0 0.010642032182541446 0.9894144 0.010585469 3.5029368E-10 0 1 2 +45 0 0 0.017220520546649732 0.9829269 0.017072704 8.5317925E-10 0 1 2 +46 0 0 0.0041847046934752278 0.99582404 0.004176236 1.5147623E-11 0 1 2 +48 0 0 0.0036583023932260659 0.9963484 0.0036516 6.698705E-12 0 1 2 +50 1 1 0.0098979546432293401 0.99015087 0.005577831 0.004271283 1 2 0 +51 1 1 0.031642692265860799 0.9688527 0.02408905 0.007058123 1 2 0 +52 1 1 0.035734869578390109 0.9648961 0.033439055 0.0016649842 1 2 0 +54 1 1 0.072408227616826942 0.9301511 0.0680363 0.0018125783 1 2 0 +56 1 1 0.093647948756734614 0.9106033 0.08498881 0.004407885 1 2 0 +60 1 1 0.042488796242443831 0.9584012 0.03626591 0.0053329817 1 2 0 +63 1 1 0.13750027638794796 0.8715341 0.12710424 0.0013615979 1 2 0 +64 1 1 0.054990287329390736 0.94649434 0.048028775 0.005476904 1 0 2 +66 1 1 0.3248465286536158 0.72263825 0.27483392 0.0025277545 1 2 0 +68 1 1 0.41495870735351531 0.66036755 0.33927298 0.00035948347 1 2 0 +69 1 1 0.021870392416744027 0.97836703 0.015369808 0.006263028 1 2 0 +70 1 2 0.95191047842075838 0.61301124 0.38600287 0.0009858216 2 1 0 +71 1 1 0.019885799499932456 0.9803106 0.01178271 0.00790666 1 0 2 +72 1 1 0.59601874075514194 0.55100095 0.44882405 0.00017507598 1 2 0 +73 1 1 0.06136439530794844 0.9404805 0.0584453 0.0010743018 1 2 0 +74 1 1 0.014404065465470026 0.9856992 0.008161224 0.006139571 1 2 0 +76 1 1 0.038601850847430071 0.9621337 0.036846306 0.0010199313 1 2 0 +77 1 1 0.29142548679139746 0.7471977 0.2520825 0.0007198269 1 2 0 +79 1 1 0.031792137584979667 0.9687079 0.030112583 0.0011796078 1 0 2 +82 1 1 0.020877603632288352 0.9793388 0.011351721 0.0093094 1 0 2 +88 1 1 0.040395996500067109 0.96040905 0.027646223 0.011944635 1 2 0 +90 1 1 0.16606971093516656 0.8469872 0.15168217 0.0013307912 1 2 0 +91 1 1 0.078325730942644556 0.9246632 0.072741956 0.0025948873 1 2 0 +92 1 1 0.023790137724984652 0.9764906 0.017208058 0.006301233 1 2 0 +93 1 1 0.031247375889365089 0.9692358 0.022428097 0.008336076 1 0 2 +95 1 1 0.026741238049869182 0.97361314 0.017711 0.0086758835 1 2 0 +96 1 1 0.046393656095039121 0.9546661 0.038534492 0.0067994706 1 2 0 +97 1 1 0.020153545641826632 0.9800482 0.014183084 0.0057688034 1 2 0 +98 1 1 0.10664007875316252 0.8988491 0.09911452 0.0020364737 1 0 2 +99 1 1 0.04247020108309911 0.958419 0.034199238 0.0073816422 1 2 0 +100 2 2 0.00061608453557407644 0.9993841 0.0006156982 5.2008225E-08 2 1 0 +102 2 2 0.03519919300651897 0.9654131 0.034584865 2.1174378E-06 2 1 0 +104 2 2 0.005589460723350561 0.99442613 0.0055736834 4.1349287E-07 2 1 0 +105 2 2 0.010776502140667386 0.98928136 0.010718442 7.400041E-08 2 1 0 +106 2 2 0.0354504447977936 0.96517056 0.034799494 2.9801784E-05 2 1 0 +108 2 2 0.025392979908311054 0.9749267 0.025072625 4.914861E-07 2 1 0 +109 2 2 0.010955341460703819 0.98910445 0.010893921 1.8230908E-06 2 1 0 +111 2 2 0.06487274446443414 0.9371867 0.06280315 1.0176921E-05 2 1 0 +112 2 2 0.065965663593841151 0.936163 0.06382359 1.3544553E-05 2 1 0 +113 2 2 0.010098914216835852 0.9899519 0.010045775 2.2422923E-06 2 1 0 +115 2 2 0.025223276429275899 0.9750922 0.024891736 1.6153237E-05 2 1 0 +117 2 2 0.040342811013684934 0.9604601 0.03953848 1.5801854E-06 2 1 0 +120 2 2 0.023179988450757825 0.9770866 0.022908697 4.5383977E-06 2 1 0 +121 2 2 0.02306701835520052 0.977197 0.022788154 1.4753952E-05 2 1 0 +122 2 2 0.010407475995621918 0.9896465 0.0103532225 3.0831316E-08 2 1 0 +123 2 2 0.32684541420395752 0.7211952 0.2786439 0.00016076972 2 1 0 +125 2 2 0.24618152097193544 0.7817803 0.21820448 1.4983249E-05 2 1 0 +128 2 2 0.0093598130805993083 0.99068385 0.009315284 7.6260625E-07 2 1 0 +129 2 1 0.72903322039689122 0.51758564 0.48237512 3.9163806E-05 1 2 0 +131 2 2 0.50473281934645287 0.60366684 0.39629033 4.2754964E-05 2 1 0 +132 2 2 0.0054106794579163063 0.99460393 0.005395588 4.55027E-07 2 1 0 +133 2 1 0.80138898584319307 0.5511017 0.4487053 0.00019299892 1 2 0 +137 2 2 0.138713849419592 0.8704771 0.12949377 2.8935885E-05 2 1 0 +138 2 2 0.41129878905894723 0.66278887 0.3366787 0.00053248147 2 1 0 +141 2 2 0.15622779242674864 0.8553643 0.14447305 0.00016275235 2 1 0 +144 2 2 0.0057010728771068496 0.99431515 0.005683327 1.4552451E-06 2 1 0 +145 2 2 0.050841354901609903 0.95042944 0.049541287 2.9361696E-05 2 1 0 +147 2 2 0.14442300556721088 0.86552155 0.13440949 6.886598E-05 2 1 0 +0 0 0 0.0038585503435748856 0.9961489 0.0038510347 4.2933796E-13 0 1 2 +1 0 0 0.012584685347583906 0.9874942 0.012506158 1.1775624E-11 0 1 2 +2 0 0 0.0067815732511543256 0.99324137 0.0067586764 3.912824E-12 0 1 2 +3 0 0 0.01748658342742293 0.9826654 0.017334426 4.7760598E-11 0 1 2 +4 0 0 0.0033483471512239845 0.99665725 0.0033429505 3.708219E-13 0 1 2 +7 0 0 0.0071277721877935993 0.99289757 0.007102496 2.3404859E-12 0 1 2 +12 0 0 0.013154037390847481 0.9869321 0.013067851 9.762894E-12 0 1 2 +13 0 0 0.0067443675996115646 0.9932783 0.0067220973 5.424766E-12 0 1 2 +14 0 0 0.00042721812913112326 0.9995729 0.00042755782 4.828237E-16 0 1 2 +15 0 0 0.00056932678870448452 0.99943084 0.0005693276 3.09127E-15 0 1 2 +16 0 0 0.0010175515838626728 0.99898297 0.0010166843 2.4305263E-14 0 1 2 +17 0 0 0.0039501022014833305 0.9960577 0.003942579 7.486923E-13 0 1 2 +19 0 0 0.0029487531496706402 0.9970556 0.0029440247 3.5918333E-13 0 1 2 +22 0 0 0.0011424985829880589 0.99885815 0.0011418032 6.2871935E-14 0 1 2 +23 0 0 0.017220520546649732 0.9829269 0.017073112 7.244564E-11 0 1 2 +24 0 0 0.031092662568657743 0.96938574 0.030613892 1.238433E-10 0 1 2 +26 0 0 0.0104513230995729 0.9896031 0.010396728 1.6004817E-11 0 1 2 +27 0 0 0.0050468627070565713 0.99496585 0.0050342605 6.6865106E-13 0 1 2 +29 0 0 0.01853939933967726 0.9816314 0.018369062 4.4390335E-11 0 1 2 +30 0 0 0.021338436107205122 0.9788876 0.021112558 5.1277774E-11 0 1 2 +33 0 0 0.00067614520641256679 0.9993241 0.00067609065 2.6482218E-15 0 1 2 +34 0 0 0.013948411370436672 0.9861484 0.013851408 9.076305E-12 0 1 2 +36 0 0 0.0020990332282393507 0.99790317 0.0020970956 4.3523463E-14 0 1 2 +38 0 0 0.012625670281340878 0.9874537 0.012546183 3.308179E-11 0 1 2 +39 0 0 0.0066604799355853899 0.99336165 0.0066381013 1.618544E-12 0 1 2 +42 0 0 0.0083122248773632463 0.9917222 0.008277643 1.1829878E-11 0 1 2 +43 0 0 0.0088872673127327953 0.9911521 0.008847519 2.9092131E-11 0 1 2 +47 0 0 0.010153766700247622 0.9898976 0.010102069 1.2726585E-11 0 1 2 +49 0 0 0.0062799535357469924 0.9937397 0.0062599946 1.7402571E-12 0 1 2 +53 1 1 0.10758208894132112 0.8980028 0.09905121 0.002945936 1 2 0 +55 1 1 0.13393617304519093 0.8746459 0.12361274 0.0017413824 1 2 0 +57 1 1 0.033590061102994395 0.9669678 0.029580938 0.0034512568 1 0 2 +58 1 1 0.014270074236615943 0.98583126 0.010978767 0.0031899887 1 2 0 +59 1 1 0.089015493070148541 0.9148314 0.07741098 0.007757726 1 2 0 +61 1 1 0.037545657530155756 0.96315044 0.028075015 0.00877447 1 2 0 +62 1 1 0.010677816776638928 0.989379 0.0066522686 0.0039686747 1 2 0 +65 1 1 0.012958259643497203 0.98712534 0.0099781845 0.0028966176 1 0 2 +67 1 1 0.011354953209760932 0.9887093 0.007062584 0.004228181 1 0 2 +75 1 1 0.01295614627421747 0.9871274 0.0075551546 0.0053175176 1 0 2 +78 1 1 0.11792734146238608 0.8887606 0.10868493 0.002554521 1 2 0 +80 1 1 0.018943815491772591 0.9812345 0.010593426 0.008172023 1 2 0 +81 1 1 0.015758287340485826 0.9843652 0.011759792 0.0038749143 1 0 2 +83 1 2 2.0407685755540683 0.8700397 0.12992881 3.1684467E-05 2 1 0 +84 1 1 0.4388854249273707 0.64475465 0.35372564 0.0015196359 1 2 0 +85 1 1 0.051713200008336381 0.9496012 0.04277773 0.007621145 1 2 0 +86 1 1 0.024159678994046224 0.97612983 0.020365698 0.0035043894 1 2 0 +87 1 1 0.065853803263157179 0.93626773 0.062357914 0.0013744461 1 2 0 +89 1 1 0.062765369775466137 0.9391638 0.05614616 0.004689919 1 2 0 +94 1 1 0.064194939523149461 0.93782216 0.058290057 0.0038878012 1 2 0 +101 2 2 0.016411789358214388 0.98372215 0.016274536 3.2268724E-06 2 1 0 +103 2 2 0.021433855574023972 0.9787942 0.021204086 1.7062545E-06 2 1 0 +107 2 2 0.015577315727075604 0.9845434 0.015456422 2.3191242E-07 2 1 0 +110 2 2 0.31495545429492572 0.7298214 0.26993796 0.0002407024 2 1 0 +114 2 2 0.0015663845070832448 0.99843484 0.001564777 3.402143E-07 2 1 0 +116 2 2 0.083010685473052695 0.9203413 0.07964616 1.26943805E-05 2 1 0 +118 2 2 8.392686160427524E-05 0.9999161 8.364592E-05 1.03120304E-10 2 1 0 +119 2 2 0.084880344829912413 0.9186222 0.08136799 9.955974E-06 2 1 0 +124 2 2 0.030456044142172706 0.97000307 0.029992132 4.883117E-06 2 1 0 +126 2 2 0.42757367525580514 0.65208936 0.34759638 0.00031436983 2 1 0 +127 2 2 0.37320041700842538 0.6885272 0.3111871 0.00028566242 2 1 0 +130 2 2 0.023496152952157262 0.97677773 0.023221876 5.795459E-07 2 1 0 +134 2 2 0.039066339502795026 0.9616869 0.03831153 1.5742246E-06 2 1 0 +135 2 2 0.012781597390713568 0.98729974 0.012699788 5.3852443E-07 2 1 0 +136 2 2 0.0040936699979112947 0.9959147 0.004084642 8.1567725E-07 2 1 0 +139 2 2 0.1202812039121738 0.88667107 0.113290906 3.8183607E-05 2 1 0 +140 2 2 0.0054459776849391699 0.9945688 0.005430249 7.5813426E-07 2 1 0 +142 2 2 0.016411789358214388 0.98372215 0.016274536 3.2268724E-06 2 1 0 +143 2 2 0.0040821790277788331 0.99592614 0.004073402 2.7964242E-07 2 1 0 +146 2 2 0.063081732631062926 0.93886673 0.061117783 1.567671E-05 2 1 0 +148 2 2 0.013316630895305281 0.98677164 0.013223605 4.9576506E-06 2 1 0 +149 2 2 0.091196544195993295 0.9128383 0.087126315 3.555931E-05 2 1 0 diff --git a/test/BaselineOutput/Common/MulticlassLogisticRegression/net9.0/MulticlassLogisticRegression-TrainTest-iris.txt b/test/BaselineOutput/Common/MulticlassLogisticRegression/net9.0/MulticlassLogisticRegression-TrainTest-iris.txt new file mode 100644 index 0000000000..4bfb07d414 --- /dev/null +++ b/test/BaselineOutput/Common/MulticlassLogisticRegression/net9.0/MulticlassLogisticRegression-TrainTest-iris.txt @@ -0,0 +1,151 @@ +Instance Label Assigned Log-loss #1 Score #2 Score #3 Score #1 Class #2 Class #3 Class +0 0 0 0.0011177346851218202 0.9988829 0.0011171503 1.3687307E-13 0 1 2 +1 0 0 0.0070638412048995951 0.99296105 0.007039365 5.9497637E-12 0 1 2 +2 0 0 0.0023155177745434415 0.99768716 0.0023129655 1.2978034E-12 0 1 2 +3 0 0 0.007210738486354412 0.9928152 0.0071849707 1.5437577E-11 0 1 2 +4 0 0 0.00077438528059811521 0.9992259 0.0007743342 9.426268E-14 0 1 2 +5 0 0 0.00071050129341987335 0.99928975 0.0007102425 1.7328237E-13 0 1 2 +6 0 0 0.0015306259954716635 0.99847054 0.0015296628 1.957049E-12 0 1 2 +7 0 0 0.0023756208579214042 0.9976272 0.0023723885 7.484667E-13 0 1 2 +8 0 0 0.010286304162673203 0.9897664 0.01023398 4.4336264E-11 0 1 2 +9 0 0 0.0076360638527052677 0.992393 0.0076067313 3.554888E-12 0 1 2 +10 0 0 0.00078106619471182768 0.99921924 0.0007807993 3.617097E-14 0 1 2 +11 0 0 0.0034971519582287528 0.99650896 0.0034912864 2.8181131E-12 0 1 2 +12 0 0 0.0075330636303078957 0.99249524 0.007504815 4.0839554E-12 0 1 2 +13 0 0 0.0024080039370842528 0.9975949 0.002404897 1.5451208E-12 0 1 2 +14 0 0 8.1065602621708377E-05 0.99991894 8.1137376E-05 1.678323E-16 0 1 2 +15 0 0 5.149973917474053E-05 0.9999485 5.1928215E-05 6.331388E-16 0 1 2 +16 0 0 0.00015259953334155612 0.9998474 0.00015296898 7.631582E-15 0 1 2 +17 0 0 0.0010500100482185809 0.99895054 0.0010492865 2.6190896E-13 0 1 2 +18 0 0 0.0010881381562196637 0.99891245 0.001087168 7.616231E-14 0 1 2 +19 0 0 0.00051022669349199779 0.9994899 0.0005100381 8.2286717E-14 0 1 2 +20 0 0 0.0050763969104369337 0.99493647 0.005063886 1.1935419E-12 0 1 2 +21 0 0 0.0006923687670788997 0.9993079 0.0006924238 3.004497E-13 0 1 2 +22 0 0 0.00016786076751786015 0.99983215 0.00016795709 1.23758984E-14 0 1 2 +23 0 0 0.0061092646340875027 0.99390936 0.0060908427 3.61644E-11 0 1 2 +24 0 0 0.011025427720164661 0.9890351 0.010964782 2.9105506E-11 0 1 2 +25 0 0 0.015137646970205936 0.98497635 0.015024051 2.140897E-11 0 1 2 +26 0 0 0.0030746591529347551 0.99693006 0.0030700709 5.9782136E-12 0 1 2 +27 0 0 0.0016374874501750563 0.99836385 0.001636383 2.273006E-13 0 1 2 +28 0 0 0.0016126515907270726 0.99838865 0.0016114973 1.9871622E-13 0 1 2 +29 0 0 0.0073089621923645761 0.9927177 0.007282561 1.3437724E-11 0 1 2 +30 0 0 0.010531433350595738 0.9895238 0.010476542 1.9456093E-11 0 1 2 +31 0 0 0.0020809352358527357 0.9979212 0.0020785842 9.195049E-13 0 1 2 +32 0 0 0.00019075305548858101 0.99980927 0.00019106721 2.4610616E-15 0 1 2 +33 0 0 8.392686160427524E-05 0.9999161 8.4108506E-05 4.981901E-16 0 1 2 +34 0 0 0.0076360638527052677 0.992393 0.0076067313 3.554888E-12 0 1 2 +35 0 0 0.0015697276052481317 0.9984315 0.0015683514 2.6222438E-13 0 1 2 +36 0 0 0.0007562515958525627 0.99924403 0.0007558848 2.1039065E-14 0 1 2 +37 0 0 0.0076360638527052677 0.992393 0.0076067313 3.554888E-12 0 1 2 +38 0 0 0.0048599128656455657 0.9951519 0.0048477952 1.0698029E-11 0 1 2 +39 0 0 0.0023708411564531844 0.99763197 0.0023682336 5.6962177E-13 0 1 2 +40 0 0 0.00071622742267819068 0.999284 0.00071621005 1.5768455E-13 0 1 2 +41 0 0 0.058551762626255131 0.9431294 0.05687025 1.3621315E-09 0 1 2 +42 0 0 0.0023279443583759346 0.99767476 0.0023251604 2.9442065E-12 0 1 2 +43 0 0 0.0018768030088227626 0.99812496 0.0018753334 1.1481032E-11 0 1 2 +44 0 0 0.0022239959620895949 0.9977785 0.0022215513 3.5708372E-12 0 1 2 +45 0 0 0.0066480593988938394 0.993374 0.006625664 1.496481E-11 0 1 2 +46 0 0 0.00079729145767329054 0.999203 0.00079705165 9.386367E-14 0 1 2 +47 0 0 0.0034036678959821066 0.9966021 0.0033979833 3.718945E-12 0 1 2 +48 0 0 0.0007820206146574389 0.9992183 0.00078217214 4.752789E-14 0 1 2 +49 0 0 0.0023393554439589317 0.9976634 0.002336341 6.543526E-13 0 1 2 +50 1 1 0.0060668068462973812 0.99395156 0.004021413 0.0020268604 1 2 0 +51 1 1 0.023338790351455526 0.97693145 0.018544413 0.0045239176 1 2 0 +52 1 1 0.031749252061110886 0.96874946 0.030576667 0.0006738069 1 2 0 +53 1 1 0.08527589608733345 0.9182589 0.08082863 0.00091243087 1 2 0 +54 1 1 0.063995836996352209 0.9380089 0.061312668 0.0006783848 1 2 0 +55 1 1 0.089581252296333444 0.914314 0.08484148 0.0008443597 1 2 0 +56 1 1 0.084591649014165271 0.91888744 0.07807882 0.0030336843 1 2 0 +57 1 1 0.019615146540724947 0.980576 0.016975438 0.0024486359 1 0 2 +58 1 1 0.0099687494540715137 0.9900808 0.009004639 0.000914755 1 2 0 +59 1 1 0.091986676106686635 0.9121173 0.08172553 0.0061572613 1 2 0 +60 1 1 0.014522532050820143 0.9855824 0.012602701 0.0018150562 1 2 0 +61 1 1 0.044980750089437339 0.9560159 0.03734064 0.00664345 1 2 0 +62 1 1 0.0041055800490160284 0.99590284 0.00352501 0.00057204254 1 2 0 +63 1 1 0.10223677829614167 0.90281576 0.09658439 0.0005996593 1 2 0 +64 1 1 0.043846617191711838 0.95710075 0.040417988 0.002481233 1 0 2 +65 1 1 0.0080723859330818042 0.9919601 0.0044040424 0.0036358412 1 0 2 +66 1 1 0.25622962578112396 0.7739642 0.22434516 0.0016907386 1 2 0 +67 1 1 0.0047101865172178353 0.9953009 0.0024510797 0.0022478926 1 0 2 +68 1 1 0.42211441184004633 0.655659 0.34426534 7.583086E-05 1 2 0 +69 1 1 0.0088934614149071764 0.99114597 0.006183518 0.0026705042 1 2 0 +70 1 2 1.1155223505147696 0.67168206 0.32774404 0.00057389005 2 1 0 +71 1 1 0.010597212881125889 0.98945874 0.0062760743 0.004265189 1 0 2 +72 1 1 0.61569359965384529 0.54026604 0.45969322 4.071747E-05 1 2 0 +73 1 1 0.03385305522413188 0.96671355 0.0328947 0.00039175275 1 2 0 +74 1 1 0.0076422502096200624 0.9923869 0.0047220076 0.0028911587 1 2 0 +75 1 1 0.0093684167343665529 0.99067533 0.006287661 0.0030368946 1 2 0 +76 1 1 0.03126336511705375 0.9692203 0.030472467 0.00030705085 1 2 0 +77 1 1 0.36661508074661348 0.6930764 0.306667 0.00025657652 1 2 0 +78 1 1 0.12292265247256561 0.88433206 0.11432233 0.0013456341 1 2 0 +79 1 1 0.017238166929688947 0.98290956 0.016735 0.00035544956 1 0 2 +80 1 1 0.0099463546352202819 0.99010295 0.0071926937 0.0027042015 1 2 0 +81 1 1 0.0061450073331952336 0.99387383 0.003742863 0.0023831835 1 0 2 +82 1 1 0.010204286632554432 0.9898476 0.0059564677 0.0041958774 1 0 2 +83 1 2 1.7857151461693934 0.83230996 0.1676771 1.2979986E-05 2 1 0 +84 1 1 0.40605498060182649 0.66627353 0.3322761 0.0014503814 1 2 0 +85 1 1 0.073174027664291069 0.92943907 0.061049104 0.009511711 1 2 0 +86 1 1 0.025716755722897309 0.9746111 0.023932807 0.0014559692 1 2 0 +87 1 1 0.048166259464637762 0.95297533 0.046818182 0.0002067161 1 2 0 +88 1 1 0.022966868716186147 0.97729486 0.013956744 0.008748454 1 2 0 +89 1 1 0.051248761298256354 0.9500423 0.047984425 0.0019731922 1 2 0 +90 1 1 0.088418472566327622 0.91537774 0.0840664 0.00055568636 1 2 0 +91 1 1 0.054397561720220852 0.9470555 0.05160925 0.0013351744 1 2 0 +92 1 1 0.011095036734795203 0.9889663 0.008229824 0.002803784 1 2 0 +93 1 1 0.014404307343127582 0.98569894 0.011823633 0.0024773271 1 0 2 +94 1 1 0.049626081163496136 0.9515852 0.046494726 0.0019200189 1 2 0 +95 1 1 0.013591263761342247 0.9865007 0.007839985 0.0056592454 1 2 0 +96 1 1 0.025276886412139963 0.9750399 0.02084066 0.004119408 1 2 0 +97 1 1 0.011028742325276021 0.98903185 0.008096678 0.00287128 1 2 0 +98 1 1 0.081860756234682974 0.92140025 0.07797241 0.0006275772 1 0 2 +99 1 1 0.022998583716414164 0.97726387 0.018543348 0.004192653 1 2 0 +100 2 2 0.00011158612051911557 0.9998884 0.000111914385 4.4149226E-09 2 1 0 +101 2 2 0.013754351388129492 0.9863398 0.013658801 1.2718294E-06 2 1 0 +102 2 2 0.0108604345367567 0.9891983 0.010801693 1.6334951E-07 2 1 0 +103 2 2 0.025912294928075799 0.97442055 0.025578955 6.915951E-07 2 1 0 +104 2 2 0.0015639965866652286 0.9984372 0.0015626437 3.6552386E-08 2 1 0 +105 2 2 0.0026331012887548114 0.99737036 0.0026293865 2.7287625E-09 2 1 0 +106 2 2 0.030572433224491991 0.9698902 0.0300982 1.1656598E-05 2 1 0 +107 2 2 0.02430917045770865 0.9759839 0.024015648 4.4953516E-08 2 1 0 +108 2 2 0.011502241068108748 0.98856366 0.01143639 3.306771E-08 2 1 0 +109 2 2 0.0019884195930131169 0.99801356 0.0019862305 1.6385486E-07 2 1 0 +110 2 2 0.16735885171843567 0.845896 0.15400647 9.76386E-05 2 1 0 +111 2 2 0.031394486428316964 0.9690932 0.030905468 1.3492925E-06 2 1 0 +112 2 2 0.023388882544591601 0.9768825 0.023115912 1.6154493E-06 2 1 0 +113 2 2 0.0044121780254528387 0.99559754 0.0044023483 3.060237E-07 2 1 0 +114 2 2 0.00052072251991020857 0.9994794 0.0005206302 9.5912995E-08 2 1 0 +115 2 2 0.0073981283724541338 0.9926292 0.00736823 2.6115715E-06 2 1 0 +116 2 2 0.084955224726875292 0.9185534 0.081441686 4.6902787E-06 2 1 0 +117 2 2 0.010474692945882892 0.98958 0.010419856 1.4996442E-07 2 1 0 +118 2 2 8.2973107700441223E-05 0.99991703 8.340647E-05 7.1105387E-12 2 1 0 +119 2 2 0.14169071213346682 0.86788964 0.13210815 2.2323643E-06 2 1 0 +120 2 2 0.0058765485330376744 0.9941407 0.0058589224 4.5094762E-07 2 1 0 +121 2 2 0.011471250334757592 0.9885943 0.011402282 3.5103346E-06 2 1 0 +122 2 2 0.0027141415686869506 0.99728954 0.0027109254 8.6256635E-10 2 1 0 +123 2 2 0.2169470530551692 0.8049726 0.19499011 3.707824E-05 2 1 0 +124 2 2 0.018596659925850037 0.9815752 0.018423153 1.8019697E-06 2 1 0 +125 2 2 0.13272763203870352 0.8757036 0.12429407 2.2208933E-06 2 1 0 +126 2 2 0.30946110547778771 0.7338423 0.26605034 0.000107192485 2 1 0 +127 2 2 0.28074074106931274 0.7552241 0.24463597 0.0001400839 2 1 0 +128 2 2 0.0030794422196350479 0.9969253 0.0030750465 6.95391E-08 2 1 0 +129 2 2 0.55820034536861218 0.57223797 0.42775527 6.9519942E-06 2 1 0 +130 2 2 0.026192673235679385 0.9741474 0.025852356 7.694972E-08 2 1 0 +131 2 2 0.22096155747459284 0.8017475 0.19824459 7.991767E-06 2 1 0 +132 2 2 0.0015125382870987553 0.9984886 0.0015117581 3.640058E-08 2 1 0 +133 2 1 0.79880115307003641 0.550074 0.44986796 5.812435E-05 1 2 0 +134 2 2 0.10817432638581932 0.89747113 0.10252805 7.112279E-07 2 1 0 +135 2 2 0.0060406014560629908 0.9939776 0.0060224 4.84113E-08 2 1 0 +136 2 2 0.0014715287541674118 0.99852955 0.0014699488 3.6579047E-07 2 1 0 +137 2 2 0.085483046947424063 0.9180687 0.08192442 6.8092263E-06 2 1 0 +138 2 2 0.31302141577516779 0.73123425 0.26854056 0.00022535904 2 1 0 +139 2 2 0.059156884560475853 0.9425589 0.05743249 8.534121E-06 2 1 0 +140 2 2 0.0018920905793713169 0.9981097 0.0018904096 1.5676574E-07 2 1 0 +141 2 2 0.047173524415149784 0.95392185 0.04605367 2.4549607E-05 2 1 0 +142 2 2 0.013754351388129492 0.9863398 0.013658801 1.2718294E-06 2 1 0 +143 2 2 0.0020293906035801714 0.99797267 0.0020276513 7.227971E-08 2 1 0 +144 2 2 0.0010890928692860219 0.9989115 0.0010883438 1.3681672E-07 2 1 0 +145 2 2 0.014195890966723048 0.9859044 0.01409219 3.5267701E-06 2 1 0 +146 2 2 0.044859430344930312 0.9561319 0.043865435 2.8941668E-06 2 1 0 +147 2 2 0.06790543976946592 0.9343488 0.06563762 1.3560748E-05 2 1 0 +148 2 2 0.0050459042079702898 0.9949668 0.0050308527 2.5297932E-06 2 1 0 +149 2 2 0.081728475570627443 0.92152214 0.0784572 2.077074E-05 2 1 0 diff --git a/test/BaselineOutput/Common/OLS/net9.0/OLS-TrainTest-generatedRegressionDataset.txt b/test/BaselineOutput/Common/OLS/net9.0/OLS-TrainTest-generatedRegressionDataset.txt new file mode 100644 index 0000000000..bda579598a --- /dev/null +++ b/test/BaselineOutput/Common/OLS/net9.0/OLS-TrainTest-generatedRegressionDataset.txt @@ -0,0 +1,501 @@ +Instance Label Score L1-loss L2-loss +0 140.66 142.00179 1.3417816162109375 1.8003779056016356 +1 148.12 147.02498 1.0950164794921875 1.1990610903594643 +2 402.2 401.97394 0.22607421875 0.051109552383422852 +3 443.51 442.01434 1.49566650390625 2.2370182909071445 +4 329.59 331.78528 2.195281982421875 4.8192629823461175 +5 322.18 323.341 1.1610107421875 1.3479459434747696 +6 425.31 425.2812 0.02880859375 0.00082993507385253906 +7 421.76 421.10257 0.657440185546875 0.43222759757190943 +8 159.37 160.24269 0.8726959228515625 0.76159817376174033 +9 325.18 326.6215 1.441497802734375 2.0779159152880311 +10 276.02 276.638 0.618011474609375 0.38193818274885416 +11 193.96 193.80179 0.1582183837890625 0.025033056968823075 +12 472.97 470.7849 2.185089111328125 4.7746144244447351 +13 266.98 265.879 1.10101318359375 1.2122300304472446 +14 331.77 332.72855 0.95855712890625 0.91883176937699318 +15 187.16 189.4515 2.29150390625 5.2509901523590088 +16 287.02 286.95404 0.065948486328125 0.00434920284897089 +17 404.13 404.35855 0.228546142578125 0.052233339287340641 +18 471.27 470.85535 0.414642333984375 0.17192826513200998 +19 449.73 449.45874 0.271270751953125 0.073587820865213871 +20 290.11 290.4139 0.303924560546875 0.092370138503611088 +21 432.91 431.12445 1.785552978515625 3.18819943908602 +22 522.87 521.5351 1.33489990234375 1.7819577492773533 +23 324.79 322.66962 2.120391845703125 4.4960615793243051 +24 456.12 454.3604 1.75958251953125 3.0961306430399418 +25 413.49 415.69672 2.20672607421875 4.8696399666368961 +26 235.38 234.83974 0.5402679443359375 0.29188945167697966 +27 476.59 479.91553 3.325531005859375 11.059156470932066 +28 360.71 361.35294 0.6429443359375 0.41337741911411285 +29 205.51 203.76144 1.7485504150390625 3.0574285539332777 +30 237.69 235.17896 2.51104736328125 6.3053588606417179 +31 372.81 374.0409 1.23089599609375 1.515104953199625 +32 349.87 348.59286 1.277130126953125 1.6310613611713052 +33 433 430.3098 2.690185546875 7.2370982766151428 +34 537 538.4245 1.42449951171875 2.0291988588869572 +35 339.31 340.2695 0.959503173828125 0.92064634058624506 +36 279.21 279.73865 0.528656005859375 0.27947717253118753 +37 326.38 326.2619 0.11810302734375 0.01394832506775856 +38 501.43 501.935 0.5050048828125 0.25502993166446686 +39 486.78 486.93372 0.153717041015625 0.023628928698599339 +40 265.78 264.25854 1.521453857421875 2.3148218402639031 +41 638.26 639.8051 1.54510498046875 2.3873494006693363 +42 439.78 438.6396 1.140411376953125 1.3005381086841226 +43 403.97 402.8758 1.094207763671875 1.1972906300798059 +44 476.58 476.0574 0.5225830078125 0.27309300005435944 +45 324.73 325.2785 0.548492431640625 0.30084394756704569 +46 155.52 157.00883 1.48883056640625 2.2166164554655552 +47 625.32 626.4237 1.10369873046875 1.2181508876383305 +48 394.42 394.881 0.46099853515625 0.21251964941620827 +49 551.95 550.67035 1.2796630859375 1.6375376135110855 +50 378.13 379.94714 1.817138671875 3.3019929528236389 +51 607.06 608.99646 1.93646240234375 3.7498866356909275 +52 814.77 813.6352 1.13482666015625 1.2878315486013889 +53 658.94 658.819 0.12103271484375 0.014648918062448502 +54 406.3 405.58154 0.71844482421875 0.51616296544671059 +55 359.51 358.38336 1.12664794921875 1.2693356014788151 +56 381.53 382.01505 0.48504638671875 0.23526999726891518 +57 351.27 349.4452 1.824798583984375 3.3298898721113801 +58 271.04 271.41998 0.379974365234375 0.14438051823526621 +59 522.23 523.4226 1.192626953125 1.4223590493202209 +60 595.54 595.9867 0.44671630859375 0.19955546036362648 +61 276.98 276.6787 0.301300048828125 0.090781719423830509 +62 398.07 398.12448 0.054473876953125 0.002967403270304203 +63 565.44 566.5533 1.11328125 1.2393951416015625 +64 503.88 506.24603 2.36602783203125 5.598087701946497 +65 278.21 276.61646 1.593536376953125 2.5393581846728921 +66 412.4 411.6762 0.723785400390625 0.52386530581861734 +67 392.53 394.02197 1.491973876953125 2.2259860495105386 +68 284.83 284.07156 0.7584228515625 0.57520522177219391 +69 342.33 342.9749 0.644927978515625 0.41593209747225046 +70 319.81 319.52954 0.28045654296875 0.078655872493982315 +71 465.97 465.5299 0.440093994140625 0.19368272367864847 +72 605.7 605.3309 0.369140625 0.13626480102539062 +73 325.76 326.68362 0.923614501953125 0.85306374821811914 +74 231.07 230.50604 0.56396484375 0.31805634498596191 +75 329.42 328.53778 0.882232666015625 0.77833447698503733 +76 497.31 497.71027 0.4002685546875 0.16021491587162018 +77 175.58 175.5811 0.0010986328125 1.2069940567016602E-06 +78 611.68 612.034 0.35400390625 0.12531876564025879 +79 313.36 312.71466 0.64532470703125 0.41644397750496864 +80 523.49 524.75574 1.2657470703125 1.6021156460046768 +81 405.43 405.0316 0.398406982421875 0.15872812364250422 +82 544.1 542.5752 1.5247802734375 2.3249548822641373 +83 557.92 558.21576 0.2957763671875 0.087483659386634827 +84 338.75 339.5174 0.76739501953125 0.58889511600136757 +85 316.99 315.87686 1.113128662109375 1.2390554184094071 +86 381.21 381.5791 0.369110107421875 0.1362422714009881 +87 540.18 538.7048 1.4752197265625 2.1762732416391373 +88 494.21 493.4995 0.710479736328125 0.50478145573288202 +89 507.79 508.40625 0.616241455078125 0.37975353095680475 +90 315.69 314.30328 1.38671875 1.9229888916015625 +91 370.5 368.41507 2.084930419921875 4.346934855915606 +92 254.02 253.8161 0.2039031982421875 0.041576514253392816 +93 599.81 599.47473 0.33526611328125 0.11240336671471596 +94 538.19 539.17413 0.984130859375 0.96851354837417603 +95 298.6 299.36157 0.761566162109375 0.57998301927000284 +96 603.69 603.4331 0.25689697265625 0.06599605455994606 +97 274.36 275.7091 1.34912109375 1.8201277256011963 +98 164.33 163.4026 0.927398681640625 0.86006831470876932 +99 414.57 415.34775 0.777740478515625 0.60488025192171335 +100 120.26 120.95938 0.69937896728515625 0.48913093988085166 +101 210.75 208.92738 1.8226165771484375 3.3219311872962862 +102 519.61 518.44293 1.16705322265625 1.3620132245123386 +103 204.42 204.96953 0.549530029296875 0.3019832530990243 +104 339.79 340.55225 0.762237548828125 0.58100608084350824 +105 523.01 521.9336 1.076416015625 1.1586714386940002 +106 372.35 371.57144 0.778564453125 0.60616260766983032 +107 406.65 407.44156 0.79156494140625 0.62657505646348 +108 368.77 370.30692 1.53692626953125 2.3621423579752445 +109 400.39 399.01547 1.374542236328125 1.889366359449923 +110 379.08 379.21957 0.13958740234375 0.019484642893075943 +111 466.77 467.69058 0.92059326171875 0.84749195352196693 +112 195.44 197.57147 2.1314697265625 4.5431631952524185 +113 594.45 594.84906 0.3990478515625 0.15923918783664703 +114 401.35 402.406 1.055999755859375 1.1151354843750596 +115 424.78 424.1251 0.6549072265625 0.42890347540378571 +116 600.71 597.5773 3.13275146484375 9.8141317404806614 +117 256.09 256.1775 0.087493896484375 0.007655181922018528 +118 242.38 242.53516 0.1551513671875 0.024071946740150452 +119 33.74 33.090424 0.64957809448242188 0.4219517008314142 +120 102.78 102.95593 0.175933837890625 0.030952715314924717 +121 443.92 442.77652 1.14349365234375 1.307577732950449 +122 356.3 356.78772 0.48773193359375 0.23788243904709816 +123 435.75 436.07825 0.3282470703125 0.10774613916873932 +124 211.81 211.44542 0.3645782470703125 0.13291729823686182 +125 518.87 517.615 1.2550048828125 1.5750372558832169 +126 370.18 371.40073 1.220733642578125 1.4901906261220574 +127 430.06 430.13864 0.078643798828125 0.0061848470941185951 +128 609.73 608.8952 0.83477783203125 0.69685402885079384 +129 397.76 397.69064 0.069366455078125 0.0048117050901055336 +130 152.17 152.10257 0.0674285888671875 0.0045466145966202021 +131 329.71 330.2685 0.558502197265625 0.3119247043505311 +132 375.37 374.90454 0.4654541015625 0.21664752066135406 +133 321.6 321.9496 0.349578857421875 0.12220537755638361 +134 362.22 363.68634 1.466339111328125 2.1501503894105554 +135 394.11 393.75403 0.35595703125 0.12670540809631348 +136 556.66 556.70593 0.04595947265625 0.0021122731268405914 +137 363.91 364.28314 0.373138427734375 0.13923228625208139 +138 293.45 292.59732 0.852691650390625 0.72708305064588785 +139 420.39 418.90552 1.4844970703125 2.2037315517663956 +140 218.78 218.29094 0.4890594482421875 0.23917914391495287 +141 386.61 386.78595 0.17596435546875 0.030963454395532608 +142 411.14 411.90967 0.7696533203125 0.59236623346805573 +143 278.87 279.643 0.77301025390625 0.59754485264420509 +144 343.98 343.32852 0.6514892578125 0.42443825304508209 +145 384.31 384.92856 0.618560791015625 0.38261745218187571 +146 176.88 177.70947 0.8294677734375 0.68801678717136383 +147 429.99 428.6915 1.298492431640625 1.6860825950279832 +148 256.74 257.01248 0.272491455078125 0.074251593090593815 +149 185.8 187.68542 1.8854217529296875 3.5548151864204556 +150 466.23 465.05222 1.17779541015625 1.3872020281851292 +151 221.42 221.90628 0.4862823486328125 0.2364705225918442 +152 166.61 166.23682 0.3731842041015625 0.13926645019091666 +153 335.06 336.2116 1.151611328125 1.3262086510658264 +154 342.29 341.1825 1.107513427734375 1.2265859926119447 +155 253.24 253.54062 0.3006134033203125 0.09036841825582087 +156 623.91 624.6078 0.69781494140625 0.48694569244980812 +157 477.18 477.06107 0.118927001953125 0.014143631793558598 +158 302.89 302.91666 0.026641845703125 0.00070978794246912003 +159 414.89 415.88724 0.997222900390625 0.99445351306349039 +160 310.87 310.36642 0.503570556640625 0.25358330551534891 +161 465.91 465.45312 0.456878662109375 0.20873811189085245 +162 137.98 141.73787 3.75787353515625 14.121613506227732 +163 570.41 569.35547 1.05450439453125 1.1119795180857182 +164 266.5 266.7501 0.250091552734375 0.062545784749090672 +165 295.17 295.8512 0.681182861328125 0.46401009056717157 +166 53.59 53.74225 0.15224838256835938 0.023179569994681515 +167 365.17 365.33615 0.1661376953125 0.027601733803749084 +168 133.92 135.41681 1.4968109130859375 2.2404429095331579 +169 463.16 464.1571 0.997100830078125 0.9942100653424859 +170 329.91 330.22552 0.315521240234375 0.099553653039038181 +171 351.11 349.2359 1.87408447265625 3.5121926106512547 +172 631.85 631.5336 0.31634521484375 0.10007429495453835 +173 292.06 293.96432 1.904327392578125 3.6264628181234002 +174 397.85 396.8436 1.00640869140625 1.0128584541380405 +175 265.37 267.66705 2.29705810546875 5.276475939899683 +176 240.33 238.14369 2.186309814453125 4.7799506047740579 +177 582.54 581.2251 1.31488037109375 1.7289103902876377 +178 587.8 587.75916 0.04083251953125 0.0016672946512699127 +179 286.32 287.98413 1.66412353515625 2.7693071402609348 +180 201.54 202.26611 0.7261199951171875 0.5272502473089844 +181 564.53 563.8096 0.720458984375 0.51906114816665649 +182 356.48 356.24078 0.239227294921875 0.05722969863563776 +183 550.32 551.13257 0.81256103515625 0.66025543585419655 +184 357.32 358.06693 0.746917724609375 0.55788608733564615 +185 378.04 378.34265 0.302642822265625 0.091592677868902683 +186 323.63 323.67902 0.04901123046875 0.0024021007120609283 +187 416.69 415.23364 1.45635986328125 2.1209840513765812 +188 525.72 523.6831 2.036865234375 4.1488199830055237 +189 522.62 521.63336 0.98663330078125 0.97344527021050453 +190 127.23 125.7359 1.4941024780273438 2.2323422148474492 +191 354.44 355.61072 1.17071533203125 1.3705743886530399 +192 428.4 430.2401 1.840118408203125 3.3860357562080026 +193 320.63 320.91406 0.2840576171875 0.080688729882240295 +194 372 370.38074 1.6192626953125 2.6220116764307022 +195 493.81 494.49896 0.68896484375 0.47467255592346191 +196 366.61 365.94714 0.662841796875 0.43935924768447876 +197 545.18 545.6168 0.43682861328125 0.19081923738121986 +198 419.62 419.98965 0.369659423828125 0.13664808962494135 +199 277.7 279.63174 1.931732177734375 3.731589206494391 +200 520.73 522.01196 1.281982421875 1.6434789299964905 +201 473.99 474.3993 0.4093017578125 0.1675279289484024 +202 599.88 601.33307 1.45306396484375 2.1113948859274387 +203 359.97 360.0347 0.064697265625 0.0041857361793518066 +204 350.02 351.0973 1.077301025390625 1.1605774993076921 +205 373.22 373.40042 0.180419921875 0.032551348209381104 +206 409.47 408.04697 1.42303466796875 2.0250276662409306 +207 578.94 579.471 0.531005859375 0.28196722269058228 +208 451.81 452.97736 1.1673583984375 1.362725630402565 +209 338.91 339.7351 0.825103759765625 0.68079621437937021 +210 402.47 403.31552 0.84552001953125 0.71490410342812538 +211 70.4 69.94559 0.45441436767578125 0.20649241755018011 +212 581.8 583.27136 1.47137451171875 2.16494295373559 +213 451.97 452.95334 0.98333740234375 0.96695244684815407 +214 535.65 536.0236 0.37359619140625 0.13957411423325539 +215 341.29 341.91104 0.62103271484375 0.3856816329061985 +216 29.14 29.269165 0.1291656494140625 0.016683764988556504 +217 207.41 207.10332 0.3066864013671875 0.094056548783555627 +218 225.92 226.90564 0.9856414794921875 0.97148912609554827 +219 397.29 396.84818 0.44183349609375 0.1952168382704258 +220 538.11 538.17334 0.0633544921875 0.0040137916803359985 +221 160.12 160.25165 0.13165283203125 0.017332468181848526 +222 456.59 457.48657 0.896575927734375 0.80384839419275522 +223 288.84 289.46826 0.628265380859375 0.39471738878637552 +224 573.66 572.95435 0.70562744140625 0.49791008606553078 +225 330.02 328.32178 1.698211669921875 2.8839228758588433 +226 197.4 198.06418 0.6641845703125 0.44114114344120026 +227 231.72 231.38742 0.33258056640625 0.11060983315110207 +228 320.12 318.2624 1.85760498046875 3.4506962634623051 +229 144.21 144.89993 0.6899261474609375 0.47599808895029128 +230 249.61 248.84439 0.7656097412109375 0.58615827583707869 +231 469.25 470.67834 1.4283447265625 2.0401686578989029 +232 371.53 371.493 0.0369873046875 0.0013680607080459595 +233 451.7 451.58972 0.11029052734375 0.012164000421762466 +234 430.73 429.73056 0.99945068359375 0.99890166893601418 +235 188.62 188.10864 0.5113525390625 0.26148141920566559 +236 235.78 235.7861 0.006103515625 3.7252902984619141E-05 +237 396.81 397.95825 1.14825439453125 1.3184881545603275 +238 424.23 424.64514 0.415130615234375 0.1723334277048707 +239 465.54 464.63464 0.905364990234375 0.81968576554208994 +240 256.29 253.25108 3.0389251708984375 9.2350661943200976 +241 161.32 160.95505 0.364959716796875 0.1331955948844552 +242 251.1 250.76428 0.335723876953125 0.11271052155643702 +243 368.25 369.5303 1.280303955078125 1.6391782173886895 +244 643.52 644.0489 0.52886962890625 0.27970308437943459 +245 353.06 353.1834 0.1234130859375 0.01523078978061676 +246 362.88 363.16748 0.2874755859375 0.082642212510108948 +247 430.27 430.59323 0.3232421875 0.10448551177978516 +248 355.19 355.94955 0.759552001953125 0.576919243671 +249 300.71 298.93872 1.771270751953125 3.1374000767245889 +250 548.8 545.5941 3.20587158203125 10.27761260047555 +251 201.68 201.15543 0.524566650390625 0.2751701707020402 +252 632.92 632.6763 0.24371337890625 0.059396211057901382 +253 442.46 443.17712 0.717132568359375 0.51427912060171366 +254 403.58 402.75394 0.8260498046875 0.6823582798242569 +255 485.05 483.92258 1.127410888671875 1.2710553118959069 +256 444.52 445.0187 0.49871826171875 0.24871990457177162 +257 391.36 390.7004 0.659576416015625 0.43504104856401682 +258 460.31 460.48172 0.171722412109375 0.029488586820662022 +259 499.14 499.9348 0.7947998046875 0.63170672953128815 +260 519.45 518.0419 1.40814208984375 1.9828641451895237 +261 244.49 244.36624 0.1237640380859375 0.015317537123337388 +262 447.03 447.1443 0.114288330078125 0.013061822392046452 +263 245.69 246.8753 1.185302734375 1.4049425721168518 +264 446.74 446.18634 0.55364990234375 0.30652821436524391 +265 494.44 493.67627 0.76373291015625 0.58328795805573463 +266 377.57 376.3369 1.23309326171875 1.5205189920961857 +267 557.56 555.52295 2.03704833984375 4.149565938860178 +268 506.71 506.333 0.376983642578125 0.1421166667714715 +269 465.86 465.32193 0.538055419921875 0.28950363490730524 +270 347.9 347.5107 0.3892822265625 0.15154065191745758 +271 368.55 369.59088 1.0408935546875 1.0834593921899796 +272 743.72 743.1681 0.5518798828125 0.30457140505313873 +273 117.89 115.38635 2.5036468505859375 6.2682475524488837 +274 398.76 397.30658 1.45343017578125 2.1124592758715153 +275 427.84 428.4129 0.572906494140625 0.32822185102850199 +276 211.26 210.84167 0.4183197021484375 0.17499137320555747 +277 477.96 476.10217 1.857818603515625 3.451489963568747 +278 195.99 196.54834 0.5583343505859375 0.31173724704422057 +279 396.87 399.19397 2.323974609375 5.4008579850196838 +280 414.67 414.1263 0.543701171875 0.29561096429824829 +281 332.09 331.29626 0.793731689453125 0.63000999484211206 +282 430.75 430.32898 0.4210205078125 0.17725826799869537 +283 283.59 281.4593 2.130706787109375 4.5399114126339555 +284 435.84 434.71484 1.125152587890625 1.2659683460369706 +285 247.75 246.04391 1.706085205078125 2.9107267269864678 +286 389.29 390.3891 1.099090576171875 1.2080000946298242 +287 474.79 474.09232 0.69769287109375 0.48677534237504005 +288 302.89 303.53354 0.643524169921875 0.41412335727363825 +289 314.35 315.50745 1.157440185546875 1.3396677831187844 +290 480.87 480.85193 0.01806640625 0.00032639503479003906 +291 478.9 478.85992 0.040069580078125 0.0016055712476372719 +292 485.49 485.9419 0.451904296875 0.20421749353408813 +293 448.7 449.8754 1.175384521484375 1.3815287733450532 +294 468.38 467.7826 0.597412109375 0.35690122842788696 +295 239.93 237.40254 2.5274505615234375 6.3880063409451395 +296 278.47 279.38943 0.91943359375 0.84535813331604004 +297 175.46 175.68507 0.225067138671875 0.050655216909945011 +298 375.75 375.19873 0.55126953125 0.30389809608459473 +299 497.29 498.6912 1.40118408203125 1.9633168317377567 +300 400.64 399.71188 0.928131103515625 0.86142734531313181 +301 329.18 332.20535 3.025360107421875 9.152803779579699 +302 501.4 500.4878 0.912200927734375 0.83211053255945444 +303 437.39 436.89954 0.490478515625 0.24056917428970337 +304 724.39 724.3373 0.052734375 0.002780914306640625 +305 323.71 320.58093 3.129058837890625 9.7910092109814286 +306 759.8 760.6697 0.86968994140625 0.75636059418320656 +307 475.94 475.85205 0.08795166015625 0.0077354945242404938 +308 588.22 587.14154 1.07843017578125 1.1630116440355778 +309 595.04 593.98676 1.05322265625 1.1092779636383057 +310 443.31 443.4859 0.1759033203125 0.030941978096961975 +311 353.34 353.3692 0.029205322265625 0.00085295084863901138 +312 476.14 474.62802 1.511993408203125 2.2861240664497018 +313 371.69 370.55417 1.135833740234375 1.2901182854548097 +314 391.02 391.6847 0.664703369140625 0.44183056894689798 +315 0 0.028259277 0.02825927734375 0.00079858675599098206 +316 362.01 364.98386 2.973846435546875 8.8437626222148538 +317 247.3 246.50314 0.7968597412109375 0.63498544716276228 +318 364.18 364.99954 0.819549560546875 0.67166148219257593 +319 333.75 334.57715 0.8271484375 0.68417453765869141 +320 188.21 188.48788 0.2778778076171875 0.077216075966134667 +321 184.42 186.29422 1.8742218017578125 3.512707362184301 +322 636.37 638.5461 2.17608642578125 4.7353521324694157 +323 433.32 433.64526 0.32525634765625 0.10579169169068336 +324 396.5 395.4945 1.0054931640625 1.0110165029764175 +325 426.49 426.58282 0.09283447265625 0.0086182393133640289 +326 271.74 272.34143 0.6014404296875 0.36173059046268463 +327 98.05 100.397705 2.3477020263671875 5.5117048046085984 +328 478.4 477.54263 0.85736083984375 0.73506760969758034 +329 424.76 427.10132 2.34130859375 5.4817259311676025 +330 508.86 509.36484 0.504852294921875 0.25487583968788385 +331 99.39 100.003845 0.6138458251953125 0.37680669710971415 +332 435.84 435.93997 0.0999755859375 0.0099951177835464478 +333 222.87 222.88858 0.018585205078125 0.00034540984779596329 +334 441.84 440.18573 1.654266357421875 2.7365971812978387 +335 373.57 374.31915 0.7491455078125 0.5612189918756485 +336 302.57 302.43768 0.13232421875 0.017509698867797852 +337 681.23 682.58545 1.35546875 1.8372955322265625 +338 533.49 531.6862 1.80377197265625 3.2535933293402195 +339 265.55 265.84732 0.297332763671875 0.08840677235275507 +340 128.89 127.66576 1.2242431640625 1.4987713247537613 +341 403.78 403.3888 0.391204833984375 0.1530412221327424 +342 442.17 442.59857 0.428558349609375 0.18366225901991129 +343 153.13 153.21976 0.089752197265625 0.0080554569140076637 +344 194.78 194.57632 0.20367431640625 0.041483227163553238 +345 177.79 179.62369 1.8336944580078125 3.3624353653285652 +346 449.69 451.1929 1.502899169921875 2.2587059149518609 +347 178.24 178.80109 0.5610809326171875 0.3148118129465729 +348 553.2 553.2258 0.02581787109375 0.00066656246781349182 +349 455.21 455.6443 0.434295654296875 0.18861271534115076 +350 513.66 511.7057 1.95428466796875 3.8192285634577274 +351 367.79 366.9605 0.829498291015625 0.6880674147978425 +352 320.48 320.1769 0.3031005859375 0.091869965195655823 +353 523.8 524.5147 0.7147216796875 0.51082707941532135 +354 482.38 482.4885 0.108489990234375 0.011770077981054783 +355 389.01 387.73584 1.274169921875 1.6235089898109436 +356 322.72 322.60925 0.110748291015625 0.012265183962881565 +357 317.61 317.6074 0.002593994140625 6.7288056015968323E-06 +358 503.57 506.22534 2.65533447265625 7.0508011616766453 +359 686.9 687.43396 0.533935546875 0.28508716821670532 +360 537.48 535.2661 2.2138671875 4.9012079238891602 +361 451.83 452.2107 0.380706787109375 0.14493765775114298 +362 346.62 346.99646 0.37646484375 0.14172577857971191 +363 376.25 376.66333 0.413330078125 0.1708417534828186 +364 244.16 243.60957 0.5504302978515625 0.30297351279295981 +365 357.16 359.15286 1.99285888671875 3.9714865423738956 +366 480.98 481.56116 0.581146240234375 0.3377309525385499 +367 304.97 306.49902 1.529022216796875 2.3379089394584298 +368 100.54 101.34741 0.80741119384765625 0.65191283595049754 +369 503.76 504.46942 0.70941162109375 0.50326484814286232 +370 288.54 290.3653 1.825286865234375 3.3316721403971314 +371 255.38 255.20686 0.1731414794921875 0.029977971920743585 +372 458.9 459.6972 0.797210693359375 0.63554488960653543 +373 318.17 319.11954 0.94952392578125 0.90159568563103676 +374 466.28 465.5068 0.773193359375 0.5978279709815979 +375 403.98 402.99365 0.986358642578125 0.97290337178856134 +376 525.67 525.7742 0.10418701171875 0.01085493341088295 +377 456.76 457.44995 0.68994140625 0.47601914405822754 +378 285.71 284.9098 0.800201416015625 0.64032230619341135 +379 384.26 383.74847 0.51153564453125 0.26166871562600136 +380 477.25 477.34122 0.091217041015625 0.0083205485716462135 +381 134.62 134.75705 0.137054443359375 0.01878392044454813 +382 189.03 190.00754 0.9775390625 0.95558261871337891 +383 393.98 393.936 0.04400634765625 0.0019365586340427399 +384 355.24 355.11487 0.1251220703125 0.015655532479286194 +385 516.48 515.259 1.22100830078125 1.4908612705767155 +386 302.84 302.64392 0.196075439453125 0.038445577956736088 +387 463.9 464.24518 0.345184326171875 0.11915221903473139 +388 263.28 264.24988 0.969879150390625 0.94066556636244059 +389 522.8 522.6209 0.1790771484375 0.032068625092506409 +390 483.43 480.79492 2.63507080078125 6.9435981251299381 +391 532.85 532.1324 0.71759033203125 0.51493588462471962 +392 234.17 234.98314 0.813140869140625 0.66119807306677103 +393 656 656.8428 0.8427734375 0.71026706695556641 +394 574.2 574.7673 0.5672607421875 0.32178474962711334 +395 446.25 448.2099 1.95989990234375 3.8412076272070408 +396 311.71 310.74103 0.968963623046875 0.93889050278812647 +397 358.82 358.59973 0.22027587890625 0.048521462827920914 +398 278.19 275.91376 2.2762451171875 5.1812918335199356 +399 172.91 172.78441 0.1255950927734375 0.015774127328768373 +400 207.26 208.1347 0.8747100830078125 0.76511772931553423 +401 456.08 455.79633 0.283660888671875 0.080463499762117863 +402 498.98 498.8804 0.099609375 0.009922027587890625 +403 107.07 107.33472 0.26471710205078125 0.070075144118163735 +404 115.82 116.2392 0.41919708251953125 0.17572619399288669 +405 332.09 331.41794 0.67205810546875 0.4516620971262455 +406 482.79 483.75977 0.969757080078125 0.94042879436165094 +407 200.21 200.35713 0.147125244140625 0.021645837463438511 +408 330.84 329.34158 1.4984130859375 2.2452417761087418 +409 329.31 328.68643 0.623565673828125 0.38883414957672358 +410 435.98 436.166 0.18597412109375 0.034586373716592789 +411 371.85 371.68552 0.16448974609375 0.027056876569986343 +412 487.32 486.30392 1.016082763671875 1.0324241826310754 +413 352.21 351.70676 0.50323486328125 0.25324532762169838 +414 311.18 311.40182 0.221832275390625 0.04920955840498209 +415 344.17 344.3455 0.17547607421875 0.030791852623224258 +416 146.42 145.0782 1.341796875 1.8004188537597656 +417 209.91 208.80042 1.109588623046875 1.2311869123950601 +418 670.07 670.23145 0.16143798828125 0.026062224060297012 +419 278.2 277.8255 0.37451171875 0.1402590274810791 +420 428.66 426.83218 1.82781982421875 3.3409253098070621 +421 525.08 523.3948 1.68524169921875 2.8400395847856998 +422 470.31 470.58405 0.2740478515625 0.075102224946022034 +423 475.06 478.62012 3.56011962890625 12.674451772123575 +424 385.98 386.56403 0.584014892578125 0.34107339475303888 +425 395.19 395.02298 0.167022705078125 0.027896584011614323 +426 524.57 524.9633 0.393310546875 0.15469318628311157 +427 361.7 360.28702 1.412994384765625 1.9965531313791871 +428 423.32 420.52228 2.7977294921875 7.8272903114557266 +429 477.74 479.32718 1.587188720703125 2.5191680351272225 +430 425.15 423.5553 1.594696044921875 2.5430554756894708 +431 584.48 586.25934 1.77935791015625 3.1661145724356174 +432 163.64 163.02734 0.6126556396484375 0.3753469327930361 +433 297.62 296.98474 0.63525390625 0.40354752540588379 +434 360.32 360.94623 0.626220703125 0.39215236902236938 +435 302.4 303.33737 0.9373779296875 0.87867738306522369 +436 356.64 356.15085 0.489166259765625 0.23928362969309092 +437 143.74 144.35698 0.616973876953125 0.38065676484256983 +438 458.22 457.82712 0.39288330078125 0.15435728803277016 +439 215.28 214.77345 0.5065460205078125 0.2565888708923012 +440 528.99 526.6168 2.3731689453125 5.6319308429956436 +441 352.24 351.4628 0.777191162109375 0.60402610246092081 +442 557.05 559.0337 1.98370361328125 3.9350800253450871 +443 558.65 559.746 1.095947265625 1.2011004090309143 +444 318.45 319.86658 1.41656494140625 2.0066562332212925 +445 488.3 488.58832 0.288330078125 0.083134233951568604 +446 334.38 333.4239 0.95611572265625 0.91415727511048317 +447 398.83 399.73233 0.90234375 0.8142242431640625 +448 623.08 623.2612 0.18121337890625 0.032838288694620132 +449 713.95 713.7672 0.18280029296875 0.033415947109460831 +450 379.64 380.75082 1.110809326171875 1.233897359110415 +451 471.12 471.15317 0.033172607421875 0.0011004218831658363 +452 228.99 228.26628 0.723724365234375 0.52377695683389902 +453 400.13 398.36868 1.761322021484375 3.1022552633658051 +454 365.24 363.46872 1.771270751953125 3.1374000767245889 +455 285.59 283.4532 2.136810302734375 4.5659582698717713 +456 529.54 528.9008 0.63916015625 0.40852570533752441 +457 375.76 377.34375 1.583740234375 2.5082331299781799 +458 497.96 497.5227 0.437286376953125 0.19121937546879053 +459 318.93 318.36398 0.566009521484375 0.32036677841097116 +460 354.93 353.9365 0.993499755859375 0.98704176489263773 +461 216.48 215.17444 1.3055572509765625 1.704479735577479 +462 628.67 629.1071 0.4371337890625 0.19108594954013824 +463 139.13 137.67984 1.450164794921875 2.1029779324308038 +464 330.37 330.23804 0.1319580078125 0.017412915825843811 +465 275.41 275.01956 0.39044189453125 0.15244487300515175 +466 394.16 392.54666 1.61334228515625 2.6028733290731907 +467 290.7 289.9436 0.75640869140625 0.57215410843491554 +468 340.51 342.8476 2.33758544921875 5.4643057323992252 +469 223.11 222.39836 0.711639404296875 0.50643064174801111 +470 476.82 478.19156 1.371551513671875 1.8811535546556115 +471 419.49 419.15262 0.337371826171875 0.11381974909454584 +472 335.12 335.20245 0.08245849609375 0.006799403578042984 +473 570.17 572.6395 2.46954345703125 6.0986448861658573 +474 415.01 416.00067 0.99066162109375 0.98141044750809669 +475 185.79 185.64262 0.147369384765625 0.021717735566198826 +476 298.86 299.61325 0.753265380859375 0.56740873400121927 +477 317.85 317.76843 0.081573486328125 0.0066542336717247963 +478 442.16 442.8144 0.654388427734375 0.42822421435266733 +479 329.43 330.01984 0.58984375 0.3479156494140625 +480 405.09 406.36438 1.274383544921875 1.6240534195676446 +481 246.03 247.1019 1.0718994140625 1.1489683538675308 +482 421.97 422.289 0.319000244140625 0.10176115576177835 +483 212.58 213.25447 0.674468994140625 0.45490842405706644 +484 457.58 459.33908 1.75909423828125 3.0944125391542912 +485 564.95 566.10425 1.15423583984375 1.3322603739798069 +486 325.87 323.81427 2.05572509765625 4.2260056771337986 +487 401.24 401.9806 0.7406005859375 0.54848922789096832 +488 320.13 319.64825 0.48175048828125 0.23208353295922279 +489 513.84 514.40015 0.56011962890625 0.31373399868607521 +490 412.16 412.53073 0.3707275390625 0.13743890821933746 +491 393.81 394.02432 0.214324951171875 0.045935184694826603 +492 239.49 239.74184 0.2518310546875 0.063418880105018616 +493 387.51 387.4073 0.10272216796875 0.010551843792200089 +494 317.12 318.38 1.260009765625 1.5876246094703674 +495 420.22 419.91824 0.3017578125 0.091057777404785156 +496 104.85 103.64624 1.2037582397460938 1.4490338997566141 +497 599.98 600.08307 0.10308837890625 0.01062721386551857 +498 414 412.82742 1.172576904296875 1.3749365964904428 +499 344.84 344.48343 0.3565673828125 0.12714029848575592 diff --git a/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Average-TrainTest-breast-cancer.txt b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Average-TrainTest-breast-cancer.txt new file mode 100644 index 0000000000..e4876c754d --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Average-TrainTest-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +0 0 -2.6282382 0.01783493 0.025962579493493374 0 +1 0 1.967058 0.9275544 3.7869586888395053 1 +2 0 -2.8538728 0.012988443 0.018861116820822133 0 +3 0 1.9359186 0.9245108 3.727585492745515 1 +4 0 -2.598872 0.018584106 0.027063459063240772 0 +5 1 6.9540396 0.99993664 9.1411674874021938E-05 1 +6 0 -1.2390391 0.11650384 0.17870423413297337 0 +7 0 -3.165355 0.008366126 0.012120539938603158 0 +8 0 -3.2661564 0.007253257 0.010502372452463592 0 +9 0 -2.8366618 0.013307133 0.019327015132525618 0 +10 0 -3.686479 0.003994234 0.005774000494858076 0 +11 0 -3.5119026 0.005118567 0.007403495120555814 0 +12 1 -0.52653205 0.26715863 1.9042314871782364 0 +13 0 -3.0795069 0.009446287 0.01369288659743881 0 +14 1 4.862514 0.99874836 0.0018068628478097768 1 +15 1 0.542482 0.62634987 0.67495935225314374 1 +16 0 -3.0233436 0.010226566 0.014829775300981111 0 +17 0 -2.8725173 0.012651704 0.018368997125237204 0 +18 1 4.350221 0.9974033 0.003751084378918107 1 +19 0 -2.3839586 0.02508771 0.036655665924702624 0 +20 1 3.2561603 0.9877451 0.017789301494361462 1 +21 1 4.061504 0.9960844 0.0056601167865794099 1 +22 0 -3.2676227 0.007238204 0.010480497109103405 0 +23 1 ? ? ? 0 +24 0 -3.605355 0.004482318 0.0064811543858276349 0 +25 1 0.5821989 0.63951796 0.64494321245002728 1 +26 0 -3.162065 0.008405169 0.012177343276744012 0 +27 0 -2.7790647 0.014430763 0.020970869910379573 0 +28 0 -3.5119026 0.005118567 0.007403495120555814 0 +29 0 -3.5155659 0.0050920118 0.0073649873764745685 0 +30 0 -3.3487473 0.0064519695 0.0093383823813951647 0 +31 0 -3.3610756 0.0063401544 0.009176028934659259 0 +32 1 3.887834 0.99498856 0.0072481557135341945 1 +33 0 -3.2242532 0.00769682 0.011147117906249074 0 +34 0 -3.071902 0.009548386 0.013841596713531164 0 +35 0 -3.5119026 0.005118567 0.007403495120555814 0 +36 1 5.102052 0.99911046 0.001283905667980111 1 +37 0 -1.6750958 0.06609421 0.098651076406760296 0 +38 1 2.925821 0.98050755 0.028399351679665001 1 +39 1 1.0715266 0.7810219 0.35656510539960956 1 +40 0 ? ? ? 0 +41 1 1.2186815 0.81482136 0.29544429037860448 1 +42 1 5.1068616 0.99911654 0.0012751267702943204 1 +43 1 0.42116195 0.5850268 0.7734253774495935 1 +44 1 4.845871 0.9987183 0.0018502574419535854 1 +45 0 -3.6570382 0.0041649234 0.0060212621744903522 0 +46 1 2.9065769 0.97997564 0.029182206242829894 1 +47 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +48 0 -2.598872 0.018584106 0.027063459063240772 0 +49 1 3.4821236 0.99109316 0.012907423641665163 1 +50 1 1.2372468 0.81878597 0.28844172110421046 1 +51 1 -0.27761188 0.34212637 1.5473987904115223 0 +52 1 2.6244812 0.97034377 0.043432146157402055 1 +53 1 3.8044782 0.9943591 0.0081611730277908311 1 +54 1 3.7374012 0.9937959 0.0089785459962833328 1 +55 1 2.637063 0.97085613 0.042670574460369946 1 +56 1 3.9728513 0.9955586 0.0064218290963102779 1 +57 1 0.10747991 0.47396427 1.0771497778884929 1 +58 1 0.94235134 0.74786824 0.41914397773752576 1 +59 1 0.5798892 0.63875765 0.64665943932445458 1 +60 1 0.99284077 0.7612114 0.3936309361034141 1 +61 0 -3.4932575 0.0052558742 0.007602620466830308 0 +62 1 4.5376863 0.99801165 0.0028714402881172539 1 +63 1 0.113746956 0.47619477 1.070376321762053 1 +64 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +65 1 0.99380887 0.76146245 0.39315519999338733 1 +66 0 -2.8725173 0.012651704 0.018368997125237204 0 +67 1 2.3670588 0.9577341 0.062302913200746018 1 +68 1 4.8663116 0.9987551 0.0017971336834092871 1 +69 0 -3.476495 0.0053824424 0.0077861961557460285 0 +70 0 -2.387623 0.024960112 0.036466855212024696 0 +71 1 3.759193 0.9939847 0.0087044508333033663 1 +72 0 -1.9715612 0.044302348 0.065373820844271124 0 +73 1 4.6668725 0.9983459 0.0023883210642366183 1 +74 1 1.3107735 0.8338385 0.26216007025720922 1 +75 0 -2.867024 0.012750015 0.018512653628799686 0 +76 0 -3.1690195 0.008322849 0.012057579489202599 0 +77 0 -2.2688777 0.029434018 0.04310179961614817 0 +78 0 -2.5909483 0.018791484 0.027368339234242232 0 +79 0 -3.4307778 0.005743255 0.0083096505009853713 0 +80 0 -2.456492 0.02267652 0.033091942438701415 0 +81 0 -2.886521 0.012404464 0.018007779743707375 0 +82 0 -2.4415562 0.023153769 0.033796614597747764 0 +83 0 -2.2505674 0.030189801 0.044225668995439241 0 +84 1 4.710059 0.9984446 0.0022456905448983559 1 +85 1 3.671128 0.9931846 0.0098661652467754053 1 +86 1 1.266887 0.8249781 0.2775722499938223 1 +87 1 3.6418068 0.9928954 0.010286319423183029 1 +88 0 -2.8725173 0.012651704 0.018368997125237204 0 +89 0 -3.2939553 0.006973092 0.01009528393823963 0 +90 0 -3.605355 0.004482318 0.0064811543858276349 0 +91 0 -3.281627 0.00709599 0.01027384406289257 0 +92 0 -2.8725173 0.012651704 0.018368997125237204 0 +93 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +94 0 -3.3610756 0.0063401544 0.009176028934659259 0 +95 0 -3.605355 0.004482318 0.0064811543858276349 0 +96 0 -3.5259056 0.0050177965 0.0072573734492600893 0 +97 0 -2.6282382 0.01783493 0.025962579493493374 0 +98 1 4.7208366 0.99846834 0.0022114130893724142 1 +99 1 5.3583617 0.9993828 0.00089071508051595711 1 +100 1 2.7089813 0.9736241 0.038563200390741868 1 +101 1 -0.573771 0.25416794 1.9761460065986429 0 +102 0 -2.6767967 0.016660688 0.02423877534032982 0 +103 1 0.1363736 0.48425505 1.0461610139478223 1 +104 1 6.2653365 0.9998308 0.00024415003072187432 1 +105 1 0.562006 0.63284785 0.66006941806243558 1 +106 1 5.5096827 0.99950266 0.00071769010260924141 1 +107 1 3.85738 0.9947671 0.0075692582256528945 1 +108 0 -3.4984608 0.0052171918 0.0075465196665404416 0 +109 1 3.2485294 0.98761255 0.017982931957058607 1 +110 0 -2.1737278 0.033571426 0.049264983756426947 0 +111 1 2.245582 0.95013195 0.07380020841571501 1 +112 1 4.0149956 0.9958168 0.0060477013299516614 1 +113 1 5.5520215 0.9995318 0.00067562003439517089 1 +114 0 -2.021864 0.04136025 0.060939334752369463 0 +115 0 -2.8553119 0.012962139 0.018822670262350266 0 +116 0 -0.7388104 0.21214464 0.34399730656570376 0 +117 1 4.22981 0.99691796 0.0044533056122670568 1 +118 0 -3.2189114 0.007755265 0.011232092952523953 0 +119 0 -2.667191 0.01688678 0.02457052117903619 0 +120 0 -3.272812 0.0071851807 0.010403444757375221 0 +121 0 -2.4229114 0.023763312 0.034697124181860606 0 +122 1 5.9445753 0.99973255 0.00038589468251374991 1 +123 1 1.9404438 0.92496026 0.11253671871535599 1 +124 1 3.7906356 0.99424714 0.008323590009243792 1 +125 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +126 1 3.8633094 0.994811 0.0075056370844255203 1 +127 0 -3.1167967 0.008961092 0.012986396064568065 0 +128 1 2.722795 0.9741257 0.037820171669183944 1 +129 0 -3.258014 0.007337415 0.010624679097672736 0 +130 0 -2.387623 0.024960112 0.036466855212024696 0 +131 0 -3.3610756 0.0063401544 0.009176028934659259 0 +132 1 5.348998 0.9993745 0.0009026753062207209 1 +133 0 -3.1387622 0.008686933 0.012587346997959896 0 +134 0 -3.186499 0.008119456 0.011761712471258297 0 +135 0 -2.1075935 0.03677224 0.054051123985110194 0 +136 0 -3.0233436 0.010226566 0.014829775300981111 0 +137 0 -3.4184494 0.005844605 0.0084567201121291263 0 +138 0 -2.8276231 0.013477576 0.019576250551719458 0 +139 0 ? ? ? 0 +140 0 -3.4184494 0.005844605 0.0084567201121291263 0 +141 0 -3.662728 0.004131378 0.0059726649365848599 0 +142 1 2.1097355 0.94010186 0.089111010678209784 1 +143 0 -2.8553119 0.012962139 0.018822670262350266 0 +144 0 -3.5119026 0.005118567 0.007403495120555814 0 +145 0 ? ? ? 0 +146 1 0.6787464 0.67063606 0.57639804079203916 1 +147 0 -3.4121335 0.005897214 0.0085330671038544147 0 +148 0 -1.1527307 0.12979358 0.20057044160696927 0 +149 1 6.111663 0.9997893 0.0003040113644898545 1 +150 0 -3.686479 0.003994234 0.005774000494858076 0 +151 1 2.2114978 0.947776 0.077381936217416319 1 +152 1 5.4552736 0.9994625 0.0007756782606498857 1 +153 0 -2.6483219 0.017339721 0.02523535479863313 0 +154 0 -3.907008 0.0029188697 0.0042171965960924847 0 +155 1 2.032926 0.933623 0.099087967767239785 1 +156 0 -3.3737788 0.006226954 0.0090116826050371697 0 +157 0 -3.3610756 0.0063401544 0.009176028934659259 0 +158 0 ? ? ? 0 +159 1 6.827044 0.99992406 0.00010955710833393459 1 +160 1 5.0816894 0.99908423 0.0013217760349626413 1 +161 0 -2.743657 0.015167387 0.022049557331495901 0 +162 0 -3.1167967 0.008961092 0.012986396064568065 0 +163 0 -2.7306023 0.015448223 0.022461016086950968 0 +164 0 ? ? ? 0 +165 0 -2.480733 0.021922318 0.031979042180753182 0 +166 1 4.648345 0.9983016 0.0024523198959080926 1 +167 1 4.109608 0.99634326 0.0052852370974816138 1 +168 0 -3.1167967 0.008961092 0.012986396064568065 0 +169 0 -3.8513095 0.0031596038 0.0045655612235389782 0 +170 0 -3.4184494 0.005844605 0.0084567201121291263 0 +171 0 -3.605355 0.004482318 0.0064811543858276349 0 +172 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +173 1 7.9525175 0.99998474 2.2013947263955502E-05 1 +174 1 3.1312795 0.9853891 0.021234562036515473 1 +175 1 4.763921 0.99855953 0.0020796505539417563 1 +176 0 -3.3610756 0.0063401544 0.009176028934659259 0 +177 1 3.233459 0.98734665 0.018371403076972846 1 +178 0 -2.8725173 0.012651704 0.018368997125237204 0 +179 1 1.1010256 0.78813696 0.34348173803507348 1 +180 0 -3.686479 0.003994234 0.005774000494858076 0 +181 0 -3.907008 0.0029188697 0.0042171965960924847 0 +182 0 -2.3839586 0.02508771 0.036655665924702624 0 +183 1 5.121356 0.9991346 0.0012490485951507386 1 +184 1 3.44398 0.9905995 0.013626183096243465 1 +185 0 -3.4421997 0.005650917 0.0081756718702575455 0 +186 1 3.04242 0.9834463 0.024081816959951673 1 +187 1 6.509462 0.99988055 0.00017233690899385432 1 +188 1 4.645643 0.99829507 0.0024617950651557919 1 +189 0 -3.0272317 0.010170551 0.01474812959715496 0 +190 1 6.920554 0.99993354 9.588351406111715E-05 1 +191 1 6.1241045 0.999793 0.00029867878857969116 1 +192 0 -2.7790647 0.014430763 0.020970869910379573 0 +193 0 -3.605355 0.004482318 0.0064811543858276349 0 +194 0 -3.1167967 0.008961092 0.012986396064568065 0 +195 0 -2.8725173 0.012651704 0.018368997125237204 0 +196 0 3.1433618 0.98563534 6.1213323654068477 1 +197 0 -2.2177703 0.031590797 0.046311305256054369 0 +198 0 -3.907008 0.0029188697 0.0042171965960924847 0 +199 0 -3.2676227 0.007238204 0.010480497109103405 0 +200 1 5.503889 0.9994985 0.00072371250314079731 1 +201 1 5.4498873 0.9994584 0.00078161486532465994 1 +202 0 -3.605355 0.004482318 0.0064811543858276349 0 +203 0 -2.6282382 0.01783493 0.025962579493493374 0 +204 0 -3.605355 0.004482318 0.0064811543858276349 0 +205 1 6.758401 0.99991626 0.00012082287142878181 1 +206 1 4.085466 0.9962156 0.0054701183827030762 1 +207 0 -3.686479 0.003994234 0.005774000494858076 0 +208 0 -3.686479 0.003994234 0.005774000494858076 0 +209 0 -2.7093625 0.015916107 0.023146784876652303 0 +210 1 7.796907 0.999981 2.7431493603031304E-05 1 +211 1 5.4917245 0.9994897 0.00073635962608825482 1 +212 0 -3.605355 0.004482318 0.0064811543858276349 0 +213 1 8.098601 0.99998766 1.7800314176101756E-05 1 +214 1 7.4137125 0.99996716 4.7381998380531431E-05 1 +215 1 4.2712793 0.9970946 0.0041977482438995033 1 +216 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +217 0 -3.605355 0.004482318 0.0064811543858276349 0 +218 1 4.417726 0.99764127 0.0034069543533184928 1 +219 0 -1.9925172 0.04305317 0.063489325348443232 0 +220 0 -3.468532 0.005443625 0.0078749446344193105 0 +221 1 5.4673867 0.9994717 0.00076234249864267256 1 +222 1 -1.7791268 0.057499222 4.12031374266423 0 +223 1 2.7707758 0.97579676 0.035347403175895874 1 +224 1 4.9876533 0.99895287 0.0015114871084349578 1 +225 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +226 1 5.2382574 0.9992675 0.0010571346112712616 1 +227 1 4.2204 0.9968764 0.00451342811475471 1 +228 0 -3.686479 0.003994234 0.005774000494858076 0 +229 1 7.388439 0.9999659 4.9187876653739729E-05 1 +230 1 2.911434 0.98011124 0.028982592371165693 1 +231 1 4.693553 0.9984076 0.0022991753372010723 1 +232 0 0.40656582 0.57996076 1.2514039962203065 1 +233 1 3.2438922 0.9875313 0.018101613111991214 1 +234 0 -1.822286 0.054250713 0.080470311927337487 0 +235 0 ? ? ? 0 +236 1 6.204999 0.9998156 0.00026608169660259049 1 +237 1 3.7684085 0.99406284 0.0085910384269904105 1 +238 1 6.983903 0.9999393 8.7541825226821807E-05 1 +239 1 2.7874248 0.9763516 0.034527287984795069 1 +240 0 -1.5933332 0.07367208 0.11040509481152809 0 +241 0 -2.9242759 0.011761417 0.01706871156077493 0 +242 0 -3.3610756 0.0063401544 0.009176028934659259 0 +243 0 -2.2520065 0.030129727 0.044136305300338685 0 +244 0 -3.605355 0.004482318 0.0064811543858276349 0 +245 0 -2.2284915 0.031126035 0.045619089143885933 0 +246 1 6.6189985 0.99989784 0.00014739666111309571 1 +247 1 1.6129208 0.88537025 0.17564719037213658 1 +248 0 -2.2282157 0.03113791 0.045636770689646861 0 +249 0 ? ? ? 0 +250 0 -3.5246053 0.0050270706 0.0072708206999751654 0 +251 1 4.0957456 0.9962705 0.0053906217053153798 1 +252 0 2.0893052 0.9384388 4.0218342065612616 1 +253 1 5.1068616 0.99911654 0.0012751267702943204 1 +254 1 4.5376863 0.99801165 0.0028714402881172539 1 +255 1 2.5934644 0.969043 0.045367385678228074 1 +256 0 -3.4184494 0.005844605 0.0084567201121291263 0 +257 0 -3.2676227 0.007238204 0.010480497109103405 0 +258 0 -3.1167967 0.008961092 0.012986396064568065 0 +259 0 2.1366 0.94222474 4.1134043671793323 1 +260 1 4.888714 0.99879426 0.0017405680714774406 1 +261 1 6.609468 0.99989647 0.00014937466503139852 1 +262 1 5.5026236 0.9994976 0.00072500302081118415 1 +263 1 4.4796762 0.9978404 0.0031190068130796983 1 +264 1 3.1048436 0.98483586 0.022044795906390829 1 +265 0 -1.9122158 0.04803001 0.07101200235915496 0 +266 1 4.4806757 0.9978435 0.0031145255934802666 1 +267 1 1.2270185 0.81661 0.29228089800983142 1 +268 1 4.497992 0.997896 0.0030386055077389349 1 +269 0 -3.605355 0.004482318 0.0064811543858276349 0 +270 1 3.4164364 0.9902263 0.014169786358717157 1 +271 0 -2.6282382 0.01783493 0.025962579493493374 0 +272 1 1.2270185 0.81661 0.29228089800983142 1 +273 1 -0.23821846 0.35489044 1.4945543996556363 0 +274 0 -2.987936 0.010750926 0.015594285288421153 0 +275 0 ? ? ? 0 +276 0 -3.2676227 0.007238204 0.010480497109103405 0 +277 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +278 0 -3.605355 0.004482318 0.0064811543858276349 0 +279 1 3.3982468 0.9899719 0.014540554264121474 1 +280 0 -3.1167967 0.008961092 0.012986396064568065 0 +281 0 -3.2242532 0.00769682 0.011147117906249074 0 +282 1 2.0520544 0.93529487 0.096506825649587999 1 +283 1 3.057497 0.98379296 0.023573362737952758 1 +284 1 3.8021343 0.99434024 0.0081885006972933494 1 +285 1 7.6579223 0.9999768 3.3451013395372324E-05 1 +286 1 8.484354 0.99999285 1.0318995955417813E-05 1 +287 0 -3.186499 0.008119456 0.011761712471258297 0 +288 1 0.6828229 0.6719198 0.5736390022172525 1 +289 1 4.5403457 0.99801916 0.0028605838354375428 1 +290 0 -3.907008 0.0029188697 0.0042171965960924847 0 +291 0 -3.605355 0.004482318 0.0064811543858276349 0 +292 1 ? ? ? 0 +293 1 2.8976893 0.97972524 0.029550884507294869 1 +294 0 ? ? ? 0 +295 1 3.931272 0.9952884 0.006813506280501468 1 +296 0 0.74386615 0.69082963 1.6935260545226383 1 +297 0 ? ? ? 0 +298 0 -1.7669983 0.058444496 0.086881952110484439 0 +299 1 3.232888 0.98733646 0.018386296116372058 1 +300 1 3.6874917 0.99334085 0.0096392532165541136 1 +301 0 -3.605355 0.004482318 0.0064811543858276349 0 +302 1 8.445241 0.99999243 1.0920939664449809E-05 1 +303 0 -3.605355 0.004482318 0.0064811543858276349 0 +304 1 2.8648682 0.9787736 0.030952915067599302 1 +305 1 4.8832994 0.9987849 0.0017540850708345254 1 +306 0 -3.605355 0.004482318 0.0064811543858276349 0 +307 0 -3.605355 0.004482318 0.0064811543858276349 0 +308 1 3.667212 0.99314666 0.0099213186592463556 1 +309 0 -1.7634478 0.058723964 0.087310228827825187 0 +310 0 -3.4307778 0.005743255 0.0083096505009853713 0 +311 0 -3.907008 0.0029188697 0.0042171965960924847 0 +312 1 2.0210109 0.93256134 0.10072947510351643 1 +313 0 -3.907008 0.0029188697 0.0042171965960924847 0 +314 0 -3.8373055 0.0032231822 0.0046575790811433796 0 +315 0 ? ? ? 0 +316 1 2.0564313 0.93567187 0.095925420185486138 1 +317 1 5.004878 0.99897826 0.0014748168706894423 1 +318 0 -3.3619823 0.0063320072 0.0091642000465731073 0 +319 0 0.92237794 0.7424553 1.9571052642251676 1 +320 1 3.4172852 0.990238 0.014152765805051364 1 +321 0 ? ? ? 0 +322 0 -3.1167967 0.008961092 0.012986396064568065 0 +323 1 2.6589572 0.9717274 0.041376488221872135 1 +324 0 -3.605355 0.004482318 0.0064811543858276349 0 +325 0 -2.6274993 0.017853409 0.02598972386028614 0 +326 1 1.6477077 0.8903133 0.16761494284401648 1 +327 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +328 1 2.586014 0.96872246 0.045844699736486819 1 +329 1 3.75527 0.99395114 0.0087531577535388729 1 +330 1 2.757298 0.9753383 0.036025413948148469 1 +331 0 -2.5290952 0.0204902 0.029868167161298809 0 +332 0 -2.1309788 0.035608202 0.052308714115951165 0 +333 1 2.531636 0.96628356 0.049481479029604751 1 +334 1 3.4191432 0.99026364 0.014115425497958325 1 +335 0 -3.907008 0.0029188697 0.0042171965960924847 0 +336 1 3.0185556 0.98288256 0.024909049876108893 1 +337 0 -3.605355 0.004482318 0.0064811543858276349 0 +338 0 -3.8373055 0.0032231822 0.0046575790811433796 0 +339 1 2.9563637 0.9813235 0.027199314263839494 1 +340 1 2.9718864 0.9817252 0.026608823011608315 1 +341 0 -3.605355 0.004482318 0.0064811543858276349 0 +342 0 -3.662728 0.004131378 0.0059726649365848599 0 +343 0 -3.907008 0.0029188697 0.0042171965960924847 0 +344 1 4.9726467 0.9989302 0.0015441984357082677 1 +345 0 -3.907008 0.0029188697 0.0042171965960924847 0 +346 0 -2.0718207 0.038623974 0.056827267289045628 0 +347 0 -3.7304115 0.0037523978 0.0054237481602311755 0 +348 1 -0.23709154 0.35525873 1.4930579803941724 0 +349 1 1.4220958 0.85470086 0.22650851519015461 1 +350 0 -2.6562457 0.01714808 0.024954022611699704 0 +351 0 -3.3610756 0.0063401544 0.009176028934659259 0 +352 0 0.2263749 0.516356 1.04798256793295 1 +353 1 5.002982 0.9989755 0.0014787765228337548 1 +354 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +355 0 -2.813359 0.013750945 0.019976082434869279 0 +356 1 -0.8004532 0.19781269 2.3377931049238438 0 +357 1 6.9374976 0.99993515 9.3561595830050562E-05 1 +358 1 3.6256385 0.9927308 0.010525546765075855 1 +359 1 2.5322683 0.96631294 0.049437606733097841 1 +360 1 8.110542 0.99998784 1.7542337039908934E-05 1 +361 1 3.4760811 0.9910167 0.013018746302090966 1 +362 0 -2.1971238 0.03250481 0.047673603369312714 0 +363 0 -1.2134647 0.12031363 0.18493883521115476 0 +364 0 -3.3610756 0.0063401544 0.009176028934659259 0 +365 0 -3.5119026 0.005118567 0.007403495120555814 0 +366 1 7.500039 0.999971 4.1878383299624531E-05 1 +367 1 6.0018344 0.99975353 0.00035561795618568968 1 +368 0 -3.5155659 0.0050920118 0.0073649873764745685 0 +369 0 -3.4458642 0.0056216065 0.0081331459487476929 0 +370 0 -2.3731344 0.025468329 0.03721902289954359 0 +371 0 -3.5155659 0.0050920118 0.0073649873764745685 0 +372 0 -2.8276231 0.013477576 0.019576250551719458 0 +373 0 -2.642242 0.017488193 0.025453349974867312 0 +374 0 -3.071902 0.009548386 0.013841596713531164 0 +375 0 -3.907008 0.0029188697 0.0042171965960924847 0 +376 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +377 0 -3.8373055 0.0032231822 0.0046575790811433796 0 +378 0 -2.6471705 0.017367741 0.025276492678993612 0 +379 0 -1.5054634 0.082701266 0.12453644596818383 0 +380 0 -3.907008 0.0029188697 0.0042171965960924847 0 +381 1 4.9334445 0.9988687 0.0016330392564727193 1 +382 0 -2.468711 0.02229324 0.032526266813835657 0 +383 0 -3.662728 0.004131378 0.0059726649365848599 0 +384 0 -3.662728 0.004131378 0.0059726649365848599 0 +385 0 -2.304271 0.028024789 0.041008574621508601 0 +386 1 3.0156004 0.98281145 0.025013427923504954 1 +387 0 -1.6454856 0.06875101 0.10276113496660831 0 +388 0 -3.383042 0.006145676 0.0088936934374697074 0 +389 0 -2.306156 0.027951604 0.040899950517214151 0 +390 0 -3.6193593 0.004394011 0.0063531862626743686 0 +391 1 5.566353 0.9995413 0.00066194107405638893 1 +392 0 -3.2676227 0.007238204 0.010480497109103405 0 +393 0 -3.988132 0.0026005863 0.0037567399049512817 0 +394 0 -3.1780581 0.008217052 0.011903673434612025 0 +395 0 -3.2676227 0.007238204 0.010480497109103405 0 +396 0 -3.1167967 0.008961092 0.012986396064568065 0 +397 0 -3.17417 0.008262398 0.011969637548813626 0 +398 0 -2.9520805 0.011309012 0.016408412638121834 0 +399 0 -3.3288846 0.006636253 0.0096059987172611994 0 +400 1 5.038892 0.99902666 0.0014049221914540643 1 +401 0 -3.4184494 0.005844605 0.0084567201121291263 0 +402 0 -2.014259 0.041792735 0.061590343742435848 0 +403 0 -2.6296775 0.017798983 0.025909777898869359 0 +404 0 -3.3178203 0.0067411643 0.0097583727125332264 0 +405 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +406 0 -2.6409411 0.01752012 0.02550023240314938 0 +407 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +408 0 -2.3922443 0.024800101 0.036230118783590899 0 +409 0 -3.071902 0.009548386 0.013841596713531164 0 +410 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +411 0 ? ? ? 0 +412 1 5.2639875 0.99929386 0.0010190990861384172 1 +413 0 -2.4325175 0.023447335 0.034230244505418145 0 +414 1 3.5109744 0.99144936 0.012389014414653879 1 +415 0 -0.7411853 0.2115787 0.3429613365601929 0 +416 1 4.889503 0.9987956 0.0017385878847548334 1 +417 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +418 0 -1.6180851 0.071297325 0.10671130478079185 0 +419 0 -3.045992 0.009904517 0.014360431908447208 0 +420 0 -1.9349009 0.046571173 0.068802848473382999 0 +421 1 6.7540197 0.9999157 0.00012159685838243442 1 +422 0 -1.8499912 0.05225739 0.07743279104640885 0 +423 0 -2.387623 0.024960112 0.036466855212024696 0 +424 0 -3.4184494 0.005844605 0.0084567201121291263 0 +425 1 8.252745 0.99999005 1.4360622819808225E-05 1 +426 0 -1.5779232 0.07518712 0.11276659939436871 0 +427 1 2.6181278 0.9700817 0.043821858484804944 1 +428 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +429 0 -3.5119026 0.005118567 0.007403495120555814 0 +430 0 -3.2539573 0.007379705 0.010686142899966281 0 +431 0 -2.063484 0.03906821 0.057494068650588183 0 +432 0 -2.6502042 0.017294006 0.025168239920027664 0 +433 0 -2.7624013 0.014772932 0.02147183015669038 0 +434 0 3.1946397 0.98663545 6.2254445560107374 1 +435 1 4.410512 0.9976169 0.003442208390054935 1 +436 1 3.1104758 0.98495543 0.021869651869936176 1 +437 0 -3.17417 0.008262398 0.011969637548813626 0 +438 0 -2.5090835 0.021071333 0.030724358803994479 0 +439 0 -2.9298906 0.011668641 0.016933278112264784 0 +440 1 4.360398 0.99744064 0.0036971146742084634 1 +441 0 -1.3482482 0.10139462 0.15424040115339974 0 +442 0 -3.1075144 0.009079509 0.01315879170382844 0 +443 0 -3.6440837 0.0042423126 0.0061333824726112034 0 +444 0 -2.0138273 0.04181742 0.061627508578946923 0 +445 0 -3.662728 0.004131378 0.0059726649365848599 0 +446 0 -3.907008 0.0029188697 0.0042171965960924847 0 +447 0 -2.9298906 0.011668641 0.016933278112264784 0 +448 0 -3.988132 0.0026005863 0.0037567399049512817 0 +449 1 5.6200657 0.9995751 0.00061316244118439554 1 +450 0 -2.8308227 0.013416996 0.019487661175230445 0 +451 0 -2.9298906 0.011668641 0.016933278112264784 0 +452 0 -3.144804 0.008612992 0.012479742641450919 0 +453 1 4.1346116 0.996471 0.005100293205883108 1 +454 0 -3.20824 0.007873345 0.011403788149716482 0 +455 1 0.087671325 0.46692148 1.0987481404323234 1 +456 1 5.275332 0.99930525 0.0010026632305847121 1 +457 1 4.727125 0.998482 0.002191691002904326 1 +458 0 -2.73417 0.015370971 0.022347820309784716 0 +459 0 -2.538449 0.020223975 0.029476105673989405 0 +460 0 -2.6562457 0.01714808 0.024954022611699704 0 +461 0 -2.3667533 0.025695339 0.037555127087118959 0 +462 0 -2.4119666 0.02412839 0.035236742680571136 0 +463 0 -3.0453098 0.0099140685 0.014374349936157852 0 +464 0 -3.17417 0.008262398 0.011969637548813626 0 +465 1 5.3521986 0.99937737 0.0008985451451290745 1 +466 1 4.8486786 0.99872345 0.0018428527164660142 1 +467 1 3.9506595 0.9954164 0.0066279344472348935 1 +468 0 -3.17417 0.008262398 0.011969637548813626 0 +469 0 -3.5604606 0.004777495 0.0069089850745956295 0 +470 0 -3.3487473 0.0064519695 0.0093383823813951647 0 +471 0 -2.4119666 0.02412839 0.035236742680571136 0 +472 0 -2.766736 0.014683161 0.021340381838255186 0 +473 0 -3.17417 0.008262398 0.011969637548813626 0 +474 0 -2.9298906 0.011668641 0.016933278112264784 0 +475 0 -3.4184494 0.005844605 0.0084567201121291263 0 +476 0 -2.9784493 0.010895878 0.015805694958811272 0 +477 0 -3.17417 0.008262398 0.011969637548813626 0 +478 0 -2.7432091 0.015176941 0.022063552465086733 0 +479 1 4.4872956 0.9978637 0.0030853118299386823 1 +480 0 -3.011015 0.010406208 0.015091645303607019 0 +481 0 -1.9503381 0.045602627 0.067338023047936041 0 +482 1 7.830414 0.9999819 2.614159977229895E-05 1 +483 1 5.8164167 0.9996789 0.00046330965618021608 1 +484 0 -2.73417 0.015370971 0.022347820309784716 0 +485 0 -3.188639 0.0080948975 0.011725993118466159 0 +486 0 -3.3487473 0.0064519695 0.0093383823813951647 0 +487 1 7.107704 0.99994916 7.3352465329757743E-05 1 +488 1 0.6287323 0.65468204 0.61113369263546335 1 +489 1 -0.7941514 0.19924372 2.3273938110593537 0 +490 0 -3.907008 0.0029188697 0.0042171965960924847 0 +491 1 3.7055154 0.9935089 0.0093952392441949729 1 +492 0 -3.1044686 0.009118702 0.013215854249290929 0 +493 1 5.802666 0.99967253 0.00047251371268429004 1 +494 0 -0.1379958 0.3882735 0.70904133723977392 0 +495 0 -3.3487473 0.0064519695 0.0093383823813951647 0 +496 0 -3.988132 0.0026005863 0.0037567399049512817 0 +497 0 -2.987488 0.0107577285 0.01560420573617809 0 +498 0 -3.0233436 0.010226566 0.014829775300981111 0 +499 0 -3.0233436 0.010226566 0.014829775300981111 0 +500 0 -2.3839586 0.02508771 0.036655665924702624 0 +501 0 -3.0233436 0.010226566 0.014829775300981111 0 +502 0 -2.886521 0.012404464 0.018007779743707375 0 +503 0 -2.8725173 0.012651704 0.018368997125237204 0 +504 0 -3.907008 0.0029188697 0.0042171965960924847 0 +505 0 -3.0450852 0.009917215 0.014378935478008999 0 +506 1 5.8826528 0.9997079 0.00042150505446367822 1 +507 0 -3.1182358 0.00894287 0.012959870793939074 0 +508 0 -2.9298906 0.011668641 0.016933278112264784 0 +509 0 -3.662728 0.004131378 0.0059726649365848599 0 +510 0 -3.907008 0.0029188697 0.0042171965960924847 0 +511 0 -2.7790647 0.014430763 0.020970869910379573 0 +512 0 -2.9298906 0.011668641 0.016933278112264784 0 +513 0 -3.3487473 0.0064519695 0.0093383823813951647 0 +514 1 5.415637 0.99943125 0.00082076265275916719 1 +515 1 4.7278605 0.9984836 0.0021893657091670671 1 +516 0 -3.988132 0.0026005863 0.0037567399049512817 0 +517 0 -3.8373055 0.0032231822 0.0046575790811433796 0 +518 0 -3.0859058 0.009361216 0.013568990389018874 0 +519 1 3.3247871 0.98887575 0.016138838657669317 1 +520 0 -3.8258834 0.0032759816 0.0047340009548722363 0 +521 0 -3.0930457 0.00926719 0.01343206371020703 0 +522 1 2.340401 0.9561669 0.064665595897767519 1 +523 1 3.9049935 0.9951092 0.007073243258192888 1 +524 0 -3.2676227 0.007238204 0.010480497109103405 0 +525 0 -3.281627 0.00709599 0.01027384406289257 0 +526 0 -3.17417 0.008262398 0.011969637548813626 0 +527 0 -2.8725173 0.012651704 0.018368997125237204 0 +528 0 -2.1573474 0.034338225 0.050410123746608811 0 +529 0 -3.1044686 0.009118702 0.013215854249290929 0 +530 1 3.2930999 0.9883671 0.01688113492724818 1 +531 0 -2.6409411 0.01752012 0.02550023240314938 0 +532 0 -3.686479 0.003994234 0.005774000494858076 0 +533 0 -3.2676227 0.007238204 0.010480497109103405 0 +534 0 -3.5119026 0.005118567 0.007403495120555814 0 +535 0 -3.077091 0.0094786035 0.013739955405207033 0 +536 0 -2.6282382 0.01783493 0.025962579493493374 0 +537 0 -2.4325175 0.023447335 0.034230244505418145 0 +538 0 -3.0233436 0.010226566 0.014829775300981111 0 +539 0 -2.534785 0.020327851 0.029629068517867445 0 +540 0 -2.5161407 0.020864576 0.030419682581076274 0 +541 0 -3.4184494 0.005844605 0.0084567201121291263 0 +542 0 -2.5565267 0.019719027 0.02873277394513845 0 +543 0 -3.0233436 0.010226566 0.014829775300981111 0 +544 0 -2.9836383 0.010816356 0.015689709947613442 0 +545 0 -2.7790647 0.014430763 0.020970869910379573 0 +546 1 6.580682 0.9998921 0.00015565269543119635 1 +547 0 -3.7438533 0.003681361 0.0053208812875972416 0 +548 0 -3.499574 0.005208953 0.0075345710924406189 0 +549 1 3.197887 0.9866964 0.019321817398433302 1 +550 0 -3.2676227 0.007238204 0.010480497109103405 0 +551 0 -3.605355 0.004482318 0.0064811543858276349 0 +552 0 -2.3178842 0.027500393 0.040230427606464114 0 +553 0 -1.1167378 0.13570663 0.21040700610639723 0 +554 0 -3.4184494 0.005844605 0.0084567201121291263 0 +555 0 -1.4463046 0.08933612 0.13500943114291578 0 +556 0 -2.3409638 0.02663305 0.038944305583168276 0 +557 0 -2.6562457 0.01714808 0.024954022611699704 0 +558 0 -3.5119026 0.005118567 0.007403495120555814 0 +559 0 -2.7790647 0.014430763 0.020970869910379573 0 +560 0 -2.6282382 0.01783493 0.025962579493493374 0 +561 0 -2.6282382 0.01783493 0.025962579493493374 0 +562 0 -3.605355 0.004482318 0.0064811543858276349 0 +563 0 -3.2676227 0.007238204 0.010480497109103405 0 +564 0 -2.743657 0.015167387 0.022049557331495901 0 +565 1 6.3099594 0.9998412 0.00022909908038257386 1 +566 0 -2.9210758 0.0118146185 0.017146381106787286 0 +567 0 -2.4974957 0.021415183 0.031231196207280109 0 +568 1 2.4063795 0.95994824 0.058971473887292125 1 +569 1 5.5644426 0.99954003 0.00066374772176177785 1 +570 1 4.5184216 0.99795634 0.0029514014404528458 1 +571 1 5.713792 0.99962825 0.00053642761249550778 1 +572 0 -3.2676227 0.007238204 0.010480497109103405 0 +573 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +574 1 3.3362236 0.98905385 0.015879029529876843 1 +575 0 -2.4325175 0.023447335 0.034230244505418145 0 +576 0 -2.7790647 0.014430763 0.020970869910379573 0 +577 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +578 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +579 0 -3.605355 0.004482318 0.0064811543858276349 0 +580 0 -2.5833437 0.018992648 0.027664145665878905 0 +581 1 4.665889 0.9983436 0.0023916802862833839 1 +582 1 4.767139 0.99856615 0.0020700917793552748 1 +583 0 -3.4184494 0.005844605 0.0084567201121291263 0 +584 0 -2.1646533 0.033994157 0.049896179290988933 0 +585 0 -3.907008 0.0029188697 0.0042171965960924847 0 +586 1 7.6839685 0.99997765 3.2247107427628552E-05 1 +587 0 -2.6502042 0.017294006 0.025168239920027664 0 +588 1 3.1214504 0.9851858 0.021532258317731347 1 +589 0 -2.9298906 0.011668641 0.016933278112264784 0 +590 1 1.9429891 0.925212 0.11214407705215272 1 +591 1 3.020261 0.9829235 0.024848946245774314 1 +592 1 2.6992416 0.97326475 0.03909578518948352 1 +593 0 -2.73417 0.015370971 0.022347820309784716 0 +594 1 2.8183753 0.97735035 0.033052272046455861 1 +595 0 -2.7790647 0.014430763 0.020970869910379573 0 +596 0 -2.8276231 0.013477576 0.019576250551719458 0 +597 0 -2.2367969 0.030770555 0.045089860846896833 0 +598 0 -3.2676227 0.007238204 0.010480497109103405 0 +599 0 -2.237171 0.030754637 0.045066166901270001 0 +600 0 -3.2676227 0.007238204 0.010480497109103405 0 +601 0 -3.8373055 0.0032231822 0.0046575790811433796 0 +602 0 -3.0233436 0.010226566 0.014829775300981111 0 +603 1 1.9442933 0.9253407 0.11194342859926387 1 +604 1 2.3615687 0.9574158 0.062782451304290096 1 +605 1 5.1624727 0.9991839 0.0011778739286676199 1 +606 0 -3.0373478 0.0100262305 0.014537795028580958 0 +607 0 -3.907008 0.0029188697 0.0042171965960924847 0 +608 1 6.1158843 0.99979055 0.00030220516721606586 1 +609 0 -2.9298906 0.011668641 0.016933278112264784 0 +610 1 3.9259372 0.99525255 0.0068654326549123252 1 +611 1 3.8206592 0.9944871 0.0079754277762202844 1 +612 1 8.576258 0.99999374 9.0291174246371028E-06 1 +613 0 -3.1385386 0.008689682 0.012591348110743092 0 +614 0 -3.5930264 0.00456152 0.0065959377597312177 0 +615 0 -2.6319022 0.017743561 0.025828375464773595 0 +616 0 -3.2676227 0.007238204 0.010480497109103405 0 +617 0 ? ? ? 0 +618 0 -3.0233436 0.010226566 0.014829775300981111 0 +619 0 -2.7790647 0.014430763 0.020970869910379573 0 +620 0 -3.2676227 0.007238204 0.010480497109103405 0 +621 0 -0.54898083 0.26093295 0.43622284392537986 0 +622 0 -1.8211399 0.054334696 0.080598429557646081 0 +623 0 -3.907008 0.0029188697 0.0042171965960924847 0 +624 0 -2.723366 0.015606087 0.022692357530866295 0 +625 0 -2.1571116 0.034349386 0.05042679828954999 0 +626 1 2.881782 0.9792693 0.030222400060270514 1 +627 0 -2.6110325 0.018270198 0.026602083913904624 0 +628 0 -3.662728 0.004131378 0.0059726649365848599 0 +629 0 -3.17417 0.008262398 0.011969637548813626 0 +630 0 -2.1074882 0.036777567 0.054059102883021654 0 +631 0 -2.7790647 0.014430763 0.020970869910379573 0 +632 0 -3.907008 0.0029188697 0.0042171965960924847 0 +633 1 2.452938 0.96242636 0.05525193107279603 1 +634 0 -3.4184494 0.005844605 0.0084567201121291263 0 +635 0 -2.831287 0.01340823 0.019474841790879004 0 +636 1 5.2248135 0.99925333 0.0010776156940094201 1 +637 0 -1.7129304 0.06283829 0.093630089825057328 0 +638 0 -3.17417 0.008262398 0.011969637548813626 0 +639 0 -2.6562457 0.01714808 0.024954022611699704 0 +640 0 -2.9005246 0.012162 0.017653627646617331 0 +641 0 -3.2676227 0.007238204 0.010480497109103405 0 +642 0 -3.2676227 0.007238204 0.010480497109103405 0 +643 0 -3.907008 0.0029188697 0.0042171965960924847 0 +644 0 -3.662728 0.004131378 0.0059726649365848599 0 +645 0 -3.2676227 0.007238204 0.010480497109103405 0 +646 0 -3.5246053 0.0050270706 0.0072708206999751654 0 +647 0 -3.633363 0.004307439 0.0062277434697571819 0 +648 1 6.0645237 0.99977463 0.00032516984380980837 1 +649 0 -3.2676227 0.007238204 0.010480497109103405 0 +650 0 -2.4117305 0.024136323 0.03524847061114763 0 +651 0 -3.295854 0.006954353 0.010068059488178942 0 +652 0 -2.6502042 0.017294006 0.025168239920027664 0 +653 0 -3.0233436 0.010226566 0.014829775300981111 0 +654 0 -3.1167967 0.008961092 0.012986396064568065 0 +655 0 -3.2676227 0.007238204 0.010480497109103405 0 +656 0 -2.7790647 0.014430763 0.020970869910379573 0 +657 0 -0.1542619 0.3827741 0.69612952756317814 0 +658 1 4.735967 0.998501 0.0021642183274458754 1 +659 0 -3.907008 0.0029188697 0.0042171965960924847 0 +660 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +661 0 -2.8725173 0.012651704 0.018368997125237204 0 +662 0 -3.36474 0.0063072937 0.0091283191384672561 0 +663 0 -3.36474 0.0063072937 0.0091283191384672561 0 +664 0 -2.9939778 0.010659607 0.015461114227987247 0 +665 0 -3.907008 0.0029188697 0.0042171965960924847 0 +666 0 -2.3115225 0.027744256 0.040592241830027159 0 +667 0 -3.1167967 0.008961092 0.012986396064568065 0 +668 1 1.3648479 0.84425753 0.24424494701870647 1 +669 1 4.8292627 0.9986876 0.0018946004886774604 1 +670 1 3.769151 0.9940691 0.0085819554390991031 1 +671 0 -2.7932918 0.014144808 0.020552344028419461 0 +672 0 -3.3610756 0.0063401544 0.009176028934659259 0 +673 0 -2.2796817 0.02899671 0.042451910631042686 0 +674 0 -3.7561812 0.0036173891 0.0052282513767424451 0 +675 0 -2.587008 0.018895455 0.027521218440060766 0 +676 0 -3.5604606 0.004777495 0.0069089850745956295 0 +677 0 -2.9298906 0.011668641 0.016933278112264784 0 +678 0 -3.907008 0.0029188697 0.0042171965960924847 0 +679 0 -3.662728 0.004131378 0.0059726649365848599 0 +680 1 8.686262 0.99999464 7.739240047104371E-06 1 +681 1 5.717595 0.9996303 0.0005335028230918109 1 +682 0 -2.4993777 0.021358965 0.031148317735992203 0 +683 0 -3.907008 0.0029188697 0.0042171965960924847 0 +684 0 -3.907008 0.0029188697 0.0042171965960924847 0 +685 0 -3.907008 0.0029188697 0.0042171965960924847 0 +686 0 -3.907008 0.0029188697 0.0042171965960924847 0 +687 0 -3.0099025 0.010422571 0.015115499694974196 0 +688 0 -3.17417 0.008262398 0.011969637548813626 0 +689 0 -3.213005 0.0078204 0.011326800902131361 0 +690 0 -3.633363 0.004307439 0.0062277434697571819 0 +691 1 3.1120288 0.9849883 0.021821547733494432 1 +692 0 -3.4184494 0.005844605 0.0084567201121291263 0 +693 0 -3.1684794 0.008329213 0.012066837489200035 0 +694 0 -3.074401 0.009514717 0.013792556275729078 0 +695 0 -3.662728 0.004131378 0.0059726649365848599 0 +696 1 4.031519 0.9959139 0.0059070398783308943 1 +697 1 2.3933108 0.959225 0.060058836685055288 1 +698 1 3.007902 0.9826249 0.025287314026062505 1 diff --git a/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer-out.txt b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer-out.txt new file mode 100644 index 0000000000..28ed905c91 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer-out.txt @@ -0,0 +1,50 @@ +maml.exe TrainTest test=%Data% tr=WeightedEnsemble{bp=AvgPer nm=3 tp=-} dout=%Output% loader=Text{col=Label:BL:0 col=Features:R4:1-9} data=%Data% out=%Output% seed=1 +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Training 3 learners for the batch 1 +Beginning training model 1 of 3 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 260 instances with missing features during training (over 10 iterations; 26 inst/iter) +Trainer 1 of 3 finished in %Time% +Beginning training model 2 of 3 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 150 instances with missing features during training (over 10 iterations; 15 inst/iter) +Trainer 2 of 3 finished in %Time% +Beginning training model 3 of 3 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 110 instances with missing features during training (over 10 iterations; 11 inst/iter) +Trainer 3 of 3 finished in %Time% +Training calibrator. +Warning: The predictor produced non-finite prediction values on 16 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3499 (239.0/(239.0+444.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 231 | 8 | 0.9665 + negative || 10 | 434 | 0.9775 + ||====================== +Precision || 0.9585 | 0.9819 | +OVERALL 0/1 ACCURACY: 0.973646 +LOG LOSS/instance: 0.112168 +Test-set entropy (prior Log-Loss/instance): 0.934003 +LOG-LOSS REDUCTION (RIG): 0.879907 +AUC: 0.996240 + +OVERALL RESULTS +--------------------------------------- +AUC: 0.996240 (0.0000) +Accuracy: 0.973646 (0.0000) +Positive precision: 0.958506 (0.0000) +Positive recall: 0.966527 (0.0000) +Negative precision: 0.981900 (0.0000) +Negative recall: 0.977477 (0.0000) +Log-loss: 0.112168 (0.0000) +Log-loss reduction: 0.879907 (0.0000) +F1 Score: 0.962500 (0.0000) +AUPRC: 0.992400 (0.0000) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer-rp.txt b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer-rp.txt new file mode 100644 index 0000000000..5a83bbedba --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer-rp.txt @@ -0,0 +1,4 @@ +WeightedEnsemble +AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC /bp /nm Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.99624 0.973646 0.958506 0.966527 0.9819 0.977477 0.112168 0.879907 0.9625 0.9924 AvgPer 3 WeightedEnsemble %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsemble{bp=AvgPer nm=3 tp=-} dout=%Output% loader=Text{col=Label:BL:0 col=Features:R4:1-9} data=%Data% out=%Output% seed=1 /bp:AvgPer;/nm:3 + diff --git a/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer.txt b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer.txt new file mode 100644 index 0000000000..f62a75d840 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-AvgPer-TrainTest-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +0 0 -3.482872 0.022534277 0.03288198330472715 0 +1 0 2.9472694 0.89578646 3.2623854205088092 1 +2 0 -4.2678914 0.011065352 0.016052908381687058 0 +3 0 2.986453 0.89910704 3.3091025690874707 1 +4 0 -3.5019512 0.022150522 0.032315689216100618 0 +5 1 10.2867985 0.999865 0.00019478350072454816 1 +6 0 -2.3605843 0.06085541 0.090580805182194468 0 +7 0 -4.9923854 0.005709156 0.0082601727532068629 0 +8 0 -4.683963 0.0075701145 0.010962914390856704 0 +9 0 -4.030381 0.013733364 0.0199503642579353 0 +10 0 -5.6011734 0.003267165 0.004721239542788672 0 +11 0 -5.476469 0.0036632803 0.0052947002111656356 0 +12 1 -0.42496967 0.27806884 1.8464860048931102 0 +13 0 -5.052911 0.005401334 0.0078135984533801803 0 +14 1 8.086931 0.9989771 0.0014764523778664522 1 +15 1 1.638257 0.7202814 0.4733674008362686 1 +16 0 -4.338106 0.010379728 0.015053041101426466 0 +17 0 -4.0520535 0.01346566 0.019558824244149512 0 +18 1 7.524542 0.99828434 0.0024773000217939069 1 +19 0 -2.9136906 0.037478253 0.055108958264505493 0 +20 1 6.2183647 0.9943106 0.00823148228811609 1 +21 1 7.010025 0.9972474 0.0039766409692087298 1 +22 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +23 1 ? ? ? 0 +24 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +25 1 1.2282019 0.6383614 0.64755468922941795 1 +26 0 -4.4920645 0.009020099 0.013072298446733502 0 +27 0 -3.7689242 0.017406764 0.025333786303575221 0 +28 0 -5.476469 0.0036632803 0.0052947002111656356 0 +29 0 -5.9523435 0.002366593 0.0034183186220779643 0 +30 0 -4.7488623 0.007134091 0.010329206580619374 0 +31 0 -5.190417 0.0047619934 0.0068865141029330135 0 +32 1 6.630597 0.99610066 0.0056365490645245509 1 +33 0 -4.8385086 0.0065725353 0.0095134621707070616 0 +34 0 -4.7176075 0.0073408713 0.010629702115580997 0 +35 0 -5.476469 0.0036632803 0.0052947002111656356 0 +36 1 8.456722 0.9992721 0.0010505084408715286 1 +37 0 -1.3033304 0.14642774 0.22841479640447576 0 +38 1 4.9043856 0.9811732 0.027420241139352593 1 +39 1 0.99844265 0.58823794 0.76552825403251523 1 +40 0 ? ? ? 0 +41 1 3.018794 0.9017767 0.14915790635460019 1 +42 1 7.673623 0.9985041 0.0021597400725719149 1 +43 1 0.637537 0.50608635 0.9825445333891023 1 +44 1 8.207812 0.9990848 0.0013210014038608894 1 +45 0 -5.777409 0.0027790966 0.0040149704628739567 0 +46 1 4.0385685 0.95915157 0.060169285608293009 1 +47 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +48 0 -3.5019512 0.022150522 0.032315689216100618 0 +49 1 5.503627 0.98907024 0.015855120398282622 1 +50 1 3.34305 0.9252361 0.11210652885516875 1 +51 1 -0.16695881 0.3281703 1.6074834186794114 0 +52 1 5.200657 0.98560375 0.02092035007777383 1 +53 1 5.2690506 0.9864706 0.019652069356953711 1 +54 1 5.9823065 0.9929389 0.010223184572271102 1 +55 1 4.7994804 0.97930324 0.030172436055860334 1 +56 1 6.316436 0.9947993 0.007522579396860297 1 +57 1 1.6828299 0.72847575 0.45704714822424003 1 +58 1 1.9818888 0.7794202 0.35952677853476617 1 +59 1 2.06083 0.7916638 0.33704016486982424 1 +60 1 1.6600094 0.72429925 0.46534220942251836 1 +61 0 -5.692307 0.0030049637 0.004341772996704873 0 +62 1 7.5410833 0.9983102 0.0024399161320377331 1 +63 1 -0.015551567 0.35961187 1.4754874576265213 0 +64 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +65 1 3.7103176 0.9455205 0.080819326205354647 1 +66 0 -4.0520535 0.01346566 0.019558824244149512 0 +67 1 3.3104677 0.92313397 0.11538806150224436 1 +68 1 8.566331 0.99934196 0.00094965672085389622 1 +69 0 -5.589958 0.0033009697 0.0047701700798309654 0 +70 0 -3.3892312 0.024513876 0.03580674391203241 0 +71 1 6.340682 0.9949136 0.0073568817512015888 1 +72 0 -3.0701742 0.032612815 0.047834669135590341 0 +73 1 7.6628513 0.9984892 0.0021812702713189381 1 +74 1 3.0784302 0.90653473 0.14156580096150426 1 +75 0 -4.5817137 0.008311308 0.012040789777465859 0 +76 0 -5.1941376 0.004745783 0.0068630154396277752 0 +77 0 -3.2187586 0.02856153 0.041805476074070584 0 +78 0 -3.914548 0.015255626 0.022178825795304873 0 +79 0 -5.6348944 0.0031675887 0.0045771175977165086 0 +80 0 -3.3903775 0.024488647 0.035769431187498965 0 +81 0 -3.9861975 0.014295496 0.020772876858829559 0 +82 0 -3.175147 0.029697139 0.043492968071745679 0 +83 0 -3.1198268 0.03120073 0.045730315913188109 0 +84 1 6.7683554 0.9965636 0.0049661956646408446 1 +85 1 5.2098436 0.9857233 0.020745342488251856 1 +86 1 2.052247 0.79035723 0.33942321314129648 1 +87 1 4.9122715 0.98130685 0.027223762657824519 1 +88 0 -4.0520535 0.01346566 0.019558824244149512 0 +89 0 -5.2829857 0.004374594 0.0063250498864140005 0 +90 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +91 0 -4.8414316 0.006554984 0.0094879737847640196 0 +92 0 -4.0520535 0.01346566 0.019558824244149512 0 +93 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +94 0 -5.190417 0.0047619934 0.0068865141029330135 0 +95 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +96 0 -5.410613 0.003891417 0.0056250799702359617 0 +97 0 -3.482872 0.022534277 0.03288198330472715 0 +98 1 8.737397 0.9994378 0.00081129825513571107 1 +99 1 9.012842 0.9995637 0.00062959385993691976 1 +100 1 4.706623 0.9774973 0.032835407454128211 1 +101 1 -1.1443181 0.165692 2.5934241335933055 0 +102 0 -3.8623738 0.015994458 0.023261653231021535 0 +103 1 1.3917551 0.6723575 0.57269956165021552 1 +104 1 10.234888 0.9998584 0.00020432985822878773 1 +105 1 2.2230587 0.815232 0.2947174500496586 1 +106 1 9.769182 0.9997825 0.00031381647486206597 1 +107 1 5.137453 0.9847542 0.022164422947398725 1 +108 0 -5.7063694 0.002966417 0.0042859952668840998 0 +109 1 5.4784746 0.98881704 0.016224495496090417 1 +110 0 -3.3888607 0.024522036 0.035818812560524833 0 +111 1 3.3876266 0.9280265 0.10776209657120317 1 +112 1 8.628673 0.9993786 0.0008967382033695237 1 +113 1 8.882472 0.9995081 0.00070986101950228368 1 +114 0 -3.109418 0.031491764 0.046163777452912752 0 +115 0 -4.9183283 0.0061096144 0.0088413465789841834 0 +116 0 -0.25753403 0.31005156 0.53543954272922578 0 +117 1 7.6684456 0.99849695 0.002170074527871256 1 +118 0 -5.8763533 0.002537693 0.003665770227128007 0 +119 0 -3.9601665 0.014637232 0.021273134285206767 0 +120 0 -5.21801 0.0046430747 0.0067141400647920081 0 +121 0 -3.390985 0.024475286 0.035749671949460417 0 +122 1 9.622346 0.99975103 0.00035923048673869361 1 +123 1 3.4594135 0.93231887 0.10110463347887823 1 +124 1 6.0973387 0.9936441 0.0091988635063169621 1 +125 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +126 1 7.136181 0.9975485 0.003541079440976193 1 +127 0 -4.621235 0.008016653 0.011612193952652979 0 +128 1 4.493952 0.9727618 0.039841504527236671 1 +129 0 -6.1793385 0.0019210462 0.0027741492798261411 0 +130 0 -3.3892312 0.024513876 0.03580674391203241 0 +131 0 -5.190417 0.0047619934 0.0068865141029330135 0 +132 1 7.395814 0.9980688 0.0027888126758143891 1 +133 0 -4.737647 0.0072076167 0.010436047725478771 0 +134 0 -5.0657125 0.0053383727 0.0077222743443433528 0 +135 0 -2.790247 0.041801494 0.061603530250817676 0 +136 0 -4.338106 0.010379728 0.015053041101426466 0 +137 0 -5.19334 0.004749254 0.0068680469780405835 0 +138 0 -4.148426 0.012336253 0.017908139491351942 0 +139 0 ? ? ? 0 +140 0 -5.19334 0.004749254 0.0068680469780405835 0 +141 0 -5.7625217 0.0028173493 0.0040703122701123903 0 +142 1 4.09754 0.96122694 0.057051011667809431 1 +143 0 -4.9183283 0.0061096144 0.0088413465789841834 0 +144 0 -5.476469 0.0036632803 0.0052947002111656356 0 +145 0 ? ? ? 0 +146 1 1.1432838 0.6201212 0.68937792751549787 1 +147 0 -5.850732 0.0025981201 0.0037531727415439252 0 +148 0 -2.192862 0.07030445 0.10516974993850199 0 +149 1 9.191527 0.99962986 0.00053410498513138422 1 +150 0 -5.6011734 0.003267165 0.004721239542788672 0 +151 1 4.351611 0.96906507 0.045334552844361685 1 +152 1 9.273411 0.99965674 0.0004953090586974708 1 +153 0 -3.917471 0.015215243 0.022119663972613043 0 +154 0 -6.331703 0.0016699883 0.0024112977743737099 0 +155 1 3.2879562 0.9216501 0.11770893420530119 1 +156 0 -5.548246 0.0034297854 0.0049566393175882141 0 +157 0 -5.190417 0.0047619934 0.0068865141029330135 0 +158 0 ? ? ? 0 +159 1 10.745465 0.9999115 0.00012770277002010949 1 +160 1 8.114793 0.999003 0.0014390944132832524 1 +161 0 -3.882413 0.015706612 0.022839691297467428 0 +162 0 -4.621235 0.008016653 0.011612193952652979 0 +163 0 -3.5052323 0.022085173 0.032219278069602737 0 +164 0 ? ? ? 0 +165 0 -3.5290694 0.021616044 0.031527348116069484 0 +166 1 6.9524126 0.99709785 0.0041930049474674319 1 +167 1 8.171877 0.999054 0.0013654142587147774 1 +168 0 -4.621235 0.008016653 0.011612193952652979 0 +169 0 -5.82137 0.0026691356 0.0038558969045012817 0 +170 0 -5.19334 0.004749254 0.0068680469780405835 0 +171 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +172 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +173 1 12.098249 0.99997455 3.6718763229178855E-05 1 +174 1 5.457567 0.9886021 0.016538121186036488 1 +175 1 6.7321186 0.99644744 0.0051343804750916197 1 +176 0 -5.190417 0.0047619934 0.0068865141029330135 0 +177 1 5.343604 0.9873569 0.018356423098867477 1 +178 0 -4.0520535 0.01346566 0.019558824244149512 0 +179 1 1.817894 0.75236696 0.41049160012054287 1 +180 0 -5.6011734 0.003267165 0.004721239542788672 0 +181 0 -6.331703 0.0016699883 0.0024112977743737099 0 +182 0 -2.9136906 0.037478253 0.055108958264505493 0 +183 1 7.498831 0.9982433 0.0025366510899867789 1 +184 1 5.510494 0.9891384 0.015755662663344983 1 +185 0 -5.031992 0.0055058105 0.0079651529807897928 0 +186 1 4.5500593 0.97409785 0.037861396865713125 1 +187 1 11.832129 0.99996746 4.6952027696279339E-05 1 +188 1 7.640403 0.9984577 0.0022268292311602884 1 +189 0 -4.9429703 0.0059733577 0.0086435749021061582 0 +190 1 10.813715 0.99991685 0.00011996288641172571 1 +191 1 10.099832 0.9998396 0.0002314212167623026 1 +192 0 -3.7689242 0.017406764 0.025333786303575221 0 +193 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +194 0 -4.621235 0.008016653 0.011612193952652979 0 +195 0 -4.0520535 0.01346566 0.019558824244149512 0 +196 0 5.527666 0.989307 6.5471876771068054 1 +197 0 -2.922678 0.03718085 0.054663255925829476 0 +198 0 -6.331703 0.0016699883 0.0024112977743737099 0 +199 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +200 1 8.758911 0.99944884 0.00079538099915120008 1 +201 1 8.602963 0.9993638 0.00091816351562154204 1 +202 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +203 0 -3.482872 0.022534277 0.03288198330472715 0 +204 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +205 1 10.099371 0.99983954 0.00023150722188518337 1 +206 1 6.829487 0.9967511 0.0046947596442636786 1 +207 0 -5.6011734 0.003267165 0.004721239542788672 0 +208 0 -5.6011734 0.003267165 0.004721239542788672 0 +209 0 -3.3692598 0.024957515 0.036463013325993401 0 +210 1 12.871542 0.9999875 1.805829135842506E-05 1 +211 1 9.103922 0.9995988 0.00057892375102128462 1 +212 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +213 1 12.87052 0.9999875 1.805829135842506E-05 1 +214 1 12.879091 0.9999876 1.7886306565083914E-05 1 +215 1 6.851302 0.99681556 0.004601503047228275 1 +216 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +217 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +218 1 6.756962 0.99652755 0.0050184007546825397 1 +219 0 -2.3909798 0.05927527 0.088155463380847759 0 +220 0 -5.40769 0.0039018646 0.0056402115997725997 0 +221 1 8.626977 0.99937767 0.00089811492069533539 1 +222 1 -2.476303 0.05504282 4.1833018519525771 0 +223 1 4.5353785 0.9737546 0.038369878966271906 1 +224 1 8.90344 0.9995175 0.00069626776733784027 1 +225 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +226 1 8.736737 0.99943745 0.00081181449340512405 1 +227 1 7.0285463 0.9972938 0.0039095566169136797 1 +228 0 -5.6011734 0.003267165 0.004721239542788672 0 +229 1 11.740766 0.9999646 5.1079751554952729E-05 1 +230 1 5.1917086 0.9854864 0.021092150358962274 1 +231 1 6.7423296 0.9964806 0.0050863996385837502 1 +232 0 1.7194624 0.7350964 1.9164606178934571 1 +233 1 5.4949408 0.98898345 0.015981712887257544 1 +234 0 -3.126328 0.031020276 0.045461617842620754 0 +235 0 ? ? ? 0 +236 1 9.951018 0.99981606 0.00026539363927387412 1 +237 1 6.384387 0.99511325 0.0070673671209627188 1 +238 1 11.298662 0.99994683 7.6706301438007131E-05 1 +239 1 5.218775 0.9858386 0.020576636425471754 1 +240 0 -1.993123 0.083318256 0.12550715258188941 0 +241 0 -4.1647773 0.012154138 0.017642145249573513 0 +242 0 -5.190417 0.0047619934 0.0068865141029330135 0 +243 0 -3.4550161 0.023106255 0.033726443005467219 0 +244 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +245 0 -2.7776203 0.0422697 0.062308648306955254 0 +246 1 10.201185 0.9998539 0.00021078013554602304 1 +247 1 3.0554304 0.90472466 0.14444930340846809 1 +248 0 -3.2038507 0.028944897 0.042374929787295484 0 +249 0 ? ? ? 0 +250 0 -5.8342986 0.0026376308 0.0038103241926819645 0 +251 1 8.333951 0.9991851 0.0011761526984772684 1 +252 0 3.7254658 0.94623464 4.2171793115221385 1 +253 1 7.673623 0.9985041 0.0021597400725719149 1 +254 1 7.5410833 0.9983102 0.0024399161320377331 1 +255 1 3.8116865 0.9501337 0.073797583784302623 1 +256 0 -5.19334 0.004749254 0.0068680469780405835 0 +257 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +258 0 -4.621235 0.008016653 0.011612193952652979 0 +259 0 3.242814 0.91859525 3.618743289344776 1 +260 1 9.219126 0.99963915 0.00052068543357584364 1 +261 1 9.322227 0.9996718 0.0004735459469822395 1 +262 1 8.625552 0.99937683 0.00089931954943311293 1 +263 1 7.8194847 0.99869186 0.0018884870944413687 1 +264 1 4.7406597 0.9781765 0.031833327029850804 1 +265 0 -2.4273925 0.057432946 0.085332839588091583 0 +266 1 7.70002 0.9985399 0.0021079827894124729 1 +267 1 1.9541712 0.77500075 0.36773038644864792 1 +268 1 8.424181 0.99925 0.0010824348145564958 1 +269 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +270 1 5.901843 0.9924 0.011006373382357104 1 +271 0 -3.482872 0.022534277 0.03288198330472715 0 +272 1 1.9541712 0.77500075 0.36773038644864792 1 +273 1 0.26800632 0.42166668 1.2458250641387991 1 +274 0 -4.4515944 0.009359387 0.013566326596317386 0 +275 0 ? ? ? 0 +276 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +277 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +278 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +279 1 6.614441 0.9960425 0.0057208076013373407 1 +280 0 -4.621235 0.008016653 0.011612193952652979 0 +281 0 -4.8385086 0.0065725353 0.0095134621707070616 0 +282 1 2.6904383 0.87155265 0.19834028187302039 1 +283 1 5.7269773 0.991084 0.012920785377970624 1 +284 1 6.3651495 0.99502635 0.0071933636614375171 1 +285 1 11.654726 0.9999617 5.5293481842253808E-05 1 +286 1 13.146536 0.9999903 1.401665413523095E-05 1 +287 0 -5.0657125 0.0053383727 0.0077222743443433528 0 +288 1 1.3048277 0.65448666 0.61156431638568187 1 +289 1 6.631444 0.9961037 0.0056321463459408586 1 +290 0 -6.331703 0.0016699883 0.0024112977743737099 0 +291 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +292 1 ? ? ? 0 +293 1 5.297947 0.9868212 0.019139422425450404 1 +294 0 ? ? ? 0 +295 1 5.721692 0.9910409 0.012983517781355288 1 +296 0 1.2067633 0.63379157 1.4492630768235966 1 +297 0 ? ? ? 0 +298 0 -2.5149822 0.053219326 0.078897837039135971 0 +299 1 6.0366135 0.993281 0.0097261698992318259 1 +300 1 7.189889 0.99766654 0.0033704082906736892 1 +301 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +302 1 13.107393 0.9999899 1.4532607192850369E-05 1 +303 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +304 1 5.193618 0.9855115 0.021055415313673605 1 +305 1 8.285159 0.99914765 0.0012302003065924249 1 +306 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +307 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +308 1 6.519493 0.9956824 0.0062424394797278623 1 +309 0 -2.4410834 0.056754254 0.084294406034601463 0 +310 0 -5.6348944 0.0031675887 0.0045771175977165086 0 +311 0 -6.331703 0.0016699883 0.0024112977743737099 0 +312 1 3.0181189 0.9017216 0.14924601953566335 1 +313 0 -6.331703 0.0016699883 0.0024112977743737099 0 +314 0 -5.8872256 0.0025124762 0.0036292980908799108 0 +315 0 ? ? ? 0 +316 1 3.2541513 0.9193725 0.12127858314360404 1 +317 1 7.468912 0.99819434 0.0026073759429630396 1 +318 0 -5.6950455 0.0029974182 0.0043308542631465951 0 +319 0 1.4018602 0.6744041 1.6188454941410018 1 +320 1 5.6245413 0.99021065 0.014192625478375142 1 +321 0 ? ? ? 0 +322 0 -4.621235 0.008016653 0.011612193952652979 0 +323 1 4.8959913 0.98102987 0.027631033936048165 1 +324 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +325 0 -4.070602 0.0132406335 0.019229786426732447 0 +326 1 3.624155 0.941286 0.087294915339676218 1 +327 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +328 1 3.361329 0.92639214 0.11030508520262793 1 +329 1 6.433284 0.99532735 0.0067570028317258763 1 +330 1 5.3615503 0.9875616 0.01805737864221952 1 +331 0 -3.2146301 0.0286672 0.041962415994304648 0 +332 0 -2.7499704 0.04331255 0.063880421371206753 0 +333 1 3.9098234 0.9542453 0.067567875521126988 1 +334 1 5.6489344 0.990426 0.013878901445224296 1 +335 0 -6.331703 0.0016699883 0.0024112977743737099 0 +336 1 4.8692474 0.9805661 0.028313232035807549 1 +337 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +338 0 -5.8872256 0.0025124762 0.0036292980908799108 0 +339 1 5.427351 0.9882843 0.017001987752258311 1 +340 1 5.5303507 0.9893331 0.015471758985912978 1 +341 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +342 0 -5.7625217 0.0028173493 0.0040703122701123903 0 +343 0 -6.331703 0.0016699883 0.0024112977743737099 0 +344 1 7.070565 0.99739623 0.0037613440240078367 1 +345 0 -6.331703 0.0016699883 0.0024112977743737099 0 +346 0 -3.024808 0.033956844 0.049840455476778289 0 +347 0 -5.833997 0.002638362 0.0038113820565777805 0 +348 1 -0.55175304 0.2552492 1.9700216427351105 0 +349 1 2.3500576 0.83220255 0.26499337935318013 1 +350 0 -3.3673954 0.02499933 0.036524884369080135 0 +351 0 -5.190417 0.0047619934 0.0068865141029330135 0 +352 0 0.03989792 0.371453 0.66990743837937095 1 +353 1 7.090433 0.9974433 0.0036932351406289404 1 +354 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +355 0 -4.3268905 0.010486355 0.015208493360958837 0 +356 1 -0.6269388 0.24231227 2.0450606496447228 0 +357 1 10.086639 0.99983764 0.00023425938852449468 1 +358 1 5.204255 0.9856507 0.020851600795979936 1 +359 1 4.969612 0.9822512 0.025836030173023942 1 +360 1 12.014021 0.99997246 3.9728539361084329E-05 1 +361 1 5.618046 0.9901525 0.014277385220010059 1 +362 0 -3.72711 0.018077724 0.02631926180179724 0 +363 0 -1.7872963 0.09898466 0.15037642436144524 0 +364 0 -5.190417 0.0047619934 0.0068865141029330135 0 +365 0 -5.476469 0.0036632803 0.0052947002111656356 0 +366 1 11.240093 0.99994385 8.1006102725794237E-05 1 +367 1 9.364464 0.99968433 0.0004554819533959377 1 +368 0 -5.9523435 0.002366593 0.0034183186220779643 0 +369 0 -5.507866 0.0035592564 0.0051440816056337706 0 +370 0 -3.6914015 0.018670755 0.027190839026833988 0 +371 0 -5.9523435 0.002366593 0.0034183186220779643 0 +372 0 -4.148426 0.012336253 0.017908139491351942 0 +373 0 -3.417016 0.023909435 0.034913082261185314 0 +374 0 -4.7176075 0.0073408713 0.010629702115580997 0 +375 0 -6.331703 0.0016699883 0.0024112977743737099 0 +376 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +377 0 -5.8872256 0.0025124762 0.0036292980908799108 0 +378 0 -3.8403473 0.01631684 0.023734390824775076 0 +379 0 -1.9753585 0.084576204 0.12748830096023772 0 +380 0 -6.331703 0.0016699883 0.0024112977743737099 0 +381 1 9.040222 0.99957454 0.00061393669231691098 1 +382 0 -3.6022816 0.020235302 0.029492784111819419 0 +383 0 -5.7625217 0.0028173493 0.0040703122701123903 0 +384 0 -5.7625217 0.0028173493 0.0040703122701123903 0 +385 0 -3.488603 0.022418326 0.032710854269247074 0 +386 1 5.1706343 0.9852062 0.021502407370847236 1 +387 0 -2.2325287 0.06795414 0.1015271492722067 0 +388 0 -5.3068285 0.00428 0.0061879865235541448 0 +389 0 -3.404365 0.024182824 0.035317218193439678 0 +390 0 -5.6937428 0.0030010054 0.0043360450985570897 0 +391 1 9.233915 0.99964404 0.00051363161678326592 1 +392 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +393 0 -6.173278 0.0019317765 0.0027896597813836964 0 +394 0 -5.3585663 0.0040816837 0.005900675385287716 0 +395 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +396 0 -4.621235 0.008016653 0.011612193952652979 0 +397 0 -4.6241584 0.007995273 0.01158109965570036 0 +398 0 -4.429922 0.009546246 0.013838480681373165 0 +399 0 -5.6446185 0.00313944 0.0045363790528858859 0 +400 1 7.713833 0.9985584 0.0020812867469309085 1 +401 0 -5.19334 0.004749254 0.0068680469780405835 0 +402 0 -2.7741146 0.04240058 0.06250581626766169 0 +403 0 -4.1333084 0.012507024 0.018157608831799624 0 +404 0 -5.1505213 0.0049393116 0.0071435771586442244 0 +405 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +406 0 -3.840701 0.016311612 0.023726722673238648 0 +407 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +408 0 -3.680035 0.018863497 0.027474226469704165 0 +409 0 -4.7176075 0.0073408713 0.010629702115580997 0 +410 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +411 0 ? ? ? 0 +412 1 8.314895 0.99917066 0.001196979721687337 1 +413 0 -3.293192 0.026720012 0.039073202769610696 0 +414 1 5.8874807 0.99229956 0.011152385793831057 1 +415 0 -1.409133 0.13466503 0.20866938414777755 0 +416 1 7.729929 0.99857956 0.0020507160792223671 1 +417 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +418 0 -1.8940392 0.09055712 0.13694506767231354 0 +419 0 -5.622347 0.0032042826 0.0046302248678838217 0 +420 0 -2.7726307 0.042456098 0.062589460707506497 0 +421 1 10.155396 0.9998476 0.00021989657667089834 1 +422 0 -2.7179632 0.044550415 0.065748344735098552 0 +423 0 -3.3892312 0.024513876 0.03580674391203241 0 +424 0 -5.19334 0.004749254 0.0068680469780405835 0 +425 1 12.06453 0.9999737 3.7922672928455309E-05 1 +426 0 -2.2380362 0.06763363 0.10103112500140572 0 +427 1 4.6389723 0.9760854 0.034920680707081007 1 +428 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +429 0 -5.476469 0.0036632803 0.0052947002111656356 0 +430 0 -5.2605605 0.0044654617 0.0064567263475559132 0 +431 0 -2.9262648 0.03706279 0.054486367564431912 0 +432 0 -3.5992837 0.020290107 0.029573485680359592 0 +433 0 -3.9921188 0.014218865 0.020660722569844799 0 +434 0 4.5476437 0.9740417 5.2676603051217787 1 +435 1 7.5759 0.99836344 0.0023629979497649372 1 +436 1 5.1114464 0.98439044 0.022697451404313559 1 +437 0 -4.6241584 0.007995273 0.01158109965570036 0 +438 0 -3.5409822 0.021385256 0.03118707632504783 0 +439 0 -4.0549765 0.013429949 0.019506602422452432 0 +440 1 8.698355 0.99941725 0.00084098225569382591 1 +441 0 -1.7145677 0.10512016 0.1602341168909554 0 +442 0 -4.921199 0.0060935845 0.0088180783132636442 0 +443 0 -5.978359 0.0023107007 0.0033374939234306631 0 +444 0 -2.7438335 0.043547317 0.064234497195446369 0 +445 0 -5.7625217 0.0028173493 0.0040703122701123903 0 +446 0 -6.331703 0.0016699883 0.0024112977743737099 0 +447 0 -4.0549765 0.013429949 0.019506602422452432 0 +448 0 -6.173278 0.0019317765 0.0027896597813836964 0 +449 1 8.28597 0.99914825 0.0012293396600255831 1 +450 0 -3.8532412 0.016127355 0.023456513967569821 0 +451 0 -4.0549765 0.013429949 0.019506602422452432 0 +452 0 -4.489523 0.009041044 0.013102790351655392 0 +453 1 5.9675255 0.99284285 0.010362708492085076 1 +454 0 -5.305497 0.004285228 0.0061955613433538137 0 +455 1 1.2593336 0.6449529 0.6327343035703008 1 +456 1 9.107019 0.99959993 0.00057728926100807164 1 +457 1 7.5455294 0.9983171 0.0024299242886964066 1 +458 0 -3.8652968 0.01595215 0.023199624712987182 0 +459 0 -3.6756172 0.018938938 0.027585161421135495 0 +460 0 -3.3673954 0.02499933 0.036524884369080135 0 +461 0 -3.7799654 0.017233716 0.025079731700697626 0 +462 0 -2.8788795 0.03865192 0.056869206933533242 0 +463 0 -4.454518 0.009334459 0.013530024095003333 0 +464 0 -4.6241584 0.007995273 0.01158109965570036 0 +465 1 8.897833 0.999515 0.00069988115098498695 1 +466 1 7.8223963 0.9986954 0.0018834069696430824 1 +467 1 6.104862 0.9936877 0.0091356031502358923 1 +468 0 -4.6241584 0.007995273 0.01158109965570036 0 +469 0 -5.855971 0.0025856486 0.0037351333082833401 0 +470 0 -4.7488623 0.007134091 0.010329206580619374 0 +471 0 -2.8788795 0.03865192 0.056869206933533242 0 +472 0 -3.3273697 0.025913613 0.037878371153099367 0 +473 0 -4.6241584 0.007995273 0.01158109965570036 0 +474 0 -4.0549765 0.013429949 0.019506602422452432 0 +475 0 -5.19334 0.004749254 0.0068680469780405835 0 +476 0 -4.4344788 0.009506654 0.013780811561449467 0 +477 0 -4.6241584 0.007995273 0.01158109965570036 0 +478 0 -3.7472515 0.017751418 0.025839914994034187 0 +479 1 7.0932884 0.99745 0.0036835794579032264 1 +480 0 -3.8965516 0.015506594 0.02254655234650613 0 +481 0 -2.716064 0.044624917 0.065860844625020301 0 +482 1 12.526953 0.99998283 2.4765714290683422E-05 1 +483 1 8.983758 0.99955183 0.00064671370472097975 1 +484 0 -3.8652968 0.01595215 0.023199624712987182 0 +485 0 -4.8072996 0.006762866 0.0097898944519384367 0 +486 0 -4.7488623 0.007134091 0.010329206580619374 0 +487 1 12.056475 0.99997354 3.8180653709006879E-05 1 +488 1 0.59977245 0.4973928 1.0075424609355903 1 +489 1 -0.7618656 0.22023608 2.1828772719853711 0 +490 0 -6.331703 0.0016699883 0.0024112977743737099 0 +491 1 5.410244 0.98810047 0.017270353764690164 1 +492 0 -4.179681 0.0119904615 0.017403124856370876 0 +493 1 8.512299 0.9993084 0.00099810252898520769 1 +494 0 -0.022896767 0.3580557 0.63947999562641378 0 +495 0 -4.7488623 0.007134091 0.010329206580619374 0 +496 0 -6.173278 0.0019317765 0.0027896597813836964 0 +497 0 -4.316434 0.01058674 0.015354861087017778 0 +498 0 -4.338106 0.010379728 0.015053041101426466 0 +499 0 -4.338106 0.010379728 0.015053041101426466 0 +500 0 -2.9136906 0.037478253 0.055108958264505493 0 +501 0 -4.338106 0.010379728 0.015053041101426466 0 +502 0 -3.9861975 0.014295496 0.020772876858829559 0 +503 0 -4.0520535 0.01346566 0.019558824244149512 0 +504 0 -6.331703 0.0016699883 0.0024112977743737099 0 +505 0 -4.5778904 0.008340376 0.012083078763414383 0 +506 1 9.572725 0.9997394 0.00037600306851729912 1 +507 0 -5.2716722 0.0044202036 0.0063911413478813421 0 +508 0 -4.0549765 0.013429949 0.019506602422452432 0 +509 0 -5.7625217 0.0028173493 0.0040703122701123903 0 +510 0 -6.331703 0.0016699883 0.0024112977743737099 0 +511 0 -3.7689242 0.017406764 0.025333786303575221 0 +512 0 -4.0549765 0.013429949 0.019506602422452432 0 +513 0 -4.7488623 0.007134091 0.010329206580619374 0 +514 1 9.177124 0.9996249 0.00054124492561566752 1 +515 1 7.4553976 0.99817175 0.0026400259790533786 1 +516 0 -6.173278 0.0019317765 0.0027896597813836964 0 +517 0 -5.8872256 0.0025124762 0.0036292980908799108 0 +518 0 -4.6517515 0.0077962438 0.011291675963435019 0 +519 1 4.6810236 0.9769728 0.033609671221477525 1 +520 0 -6.490128 0.0014436261 0.0020842170305833373 0 +521 0 -4.782583 0.006917477 0.010014487502899325 0 +522 1 4.839096 0.9800299 0.029102357379861676 1 +523 1 6.776594 0.99658954 0.0049286609408553195 1 +524 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +525 0 -4.8414316 0.006554984 0.0094879737847640196 0 +526 0 -4.6241584 0.007995273 0.01158109965570036 0 +527 0 -4.0520535 0.01346566 0.019558824244149512 0 +528 0 -2.754527 0.043139014 0.063618752144016147 0 +529 0 -4.179681 0.0119904615 0.017403124856370876 0 +530 1 5.300786 0.98685515 0.019089753637405036 1 +531 0 -3.840701 0.016311612 0.023726722673238648 0 +532 0 -5.6011734 0.003267165 0.004721239542788672 0 +533 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +534 0 -5.476469 0.0036632803 0.0052947002111656356 0 +535 0 -4.9217196 0.006090681 0.0088138639205450325 0 +536 0 -3.482872 0.022534277 0.03288198330472715 0 +537 0 -3.293192 0.026720012 0.039073202769610696 0 +538 0 -4.338106 0.010379728 0.015053041101426466 0 +539 0 -3.1997423 0.029051421 0.042533202012500597 0 +540 0 -3.4155807 0.023940299 0.034958701068267019 0 +541 0 -5.19334 0.004749254 0.0068680469780405835 0 +542 0 -3.4864955 0.0224609 0.032773686184538714 0 +543 0 -4.338106 0.010379728 0.015053041101426466 0 +544 0 -4.745201 0.007158011 0.010363964355243683 0 +545 0 -3.7689242 0.017406764 0.025333786303575221 0 +546 1 10.264343 0.99986213 0.00019891164756910478 1 +547 0 -5.6040964 0.0032584115 0.0047085695710448704 0 +548 0 -5.034915 0.005491092 0.0079438010955671743 0 +549 1 5.6715975 0.99062186 0.013593630704674639 1 +550 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +551 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +552 0 -3.3162293 0.026173824 0.038263815437940364 0 +553 0 -1.7405229 0.10289294 0.15664793384399639 0 +554 0 -5.19334 0.004749254 0.0068680469780405835 0 +555 0 -1.963068 0.08545658 0.12887642798412488 0 +556 0 -2.9261155 0.037067696 0.054493718177675114 0 +557 0 -3.3673954 0.02499933 0.036524884369080135 0 +558 0 -5.476469 0.0036632803 0.0052947002111656356 0 +559 0 -3.7689242 0.017406764 0.025333786303575221 0 +560 0 -3.482872 0.022534277 0.03288198330472715 0 +561 0 -3.482872 0.022534277 0.03288198330472715 0 +562 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +563 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +564 0 -3.882413 0.015706612 0.022839691297467428 0 +565 1 9.47172 0.999714 0.00041264538703320383 1 +566 0 -4.4315553 0.009532037 0.013817783724351083 0 +567 0 -3.6314182 0.019710157 0.028719720007977147 0 +568 1 4.041501 0.95925725 0.060010338652604994 1 +569 1 9.748816 0.9997784 0.00031975117930321329 1 +570 1 6.448406 0.99539167 0.0066637856180221805 1 +571 1 9.226857 0.9996417 0.00051698647656674178 1 +572 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +573 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +574 1 5.8652763 0.9921417 0.011381876111811766 1 +575 0 -3.293192 0.026720012 0.039073202769610696 0 +576 0 -3.7689242 0.017406764 0.025333786303575221 0 +577 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +578 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +579 0 -5.7595987 0.0028249214 0.0040812674334768451 0 +580 0 -3.5792446 0.020660186 0.030118558204333054 0 +581 1 7.14419 0.9975665 0.0035150464757207813 1 +582 1 7.248555 0.99778897 0.0031933798554606417 1 +583 0 -5.19334 0.004749254 0.0068680469780405835 0 +584 0 -2.9208136 0.037242357 0.054755423294747634 0 +585 0 -6.331703 0.0016699883 0.0024112977743737099 0 +586 1 11.944014 0.9999706 4.2394346321571357E-05 1 +587 0 -3.5992837 0.020290107 0.029573485680359592 0 +588 1 5.449993 0.98852324 0.016653203948794274 1 +589 0 -4.0549765 0.013429949 0.019506602422452432 0 +590 1 3.1383781 0.91110826 0.13430561360398563 1 +591 1 4.1688643 0.96360195 0.053490786532823395 1 +592 1 5.2628794 0.9863945 0.019763303451007887 1 +593 0 -3.8652968 0.01595215 0.023199624712987182 0 +594 1 3.7616892 0.94790655 0.07718325210685581 1 +595 0 -3.7689242 0.017406764 0.025333786303575221 0 +596 0 -4.148426 0.012336253 0.017908139491351942 0 +597 0 -2.9417696 0.03655659 0.053728166314280686 0 +598 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +599 0 -3.0197873 0.034108832 0.050067453365076842 0 +600 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +601 0 -5.8872256 0.0025124762 0.0036292980908799108 0 +602 0 -4.338106 0.010379728 0.015053041101426466 0 +603 1 3.4922361 0.9342013 0.098194641924081588 1 +604 1 4.188937 0.9642447 0.052528748937019649 1 +605 1 9.458144 0.99971044 0.0004178063517362282 1 +606 0 -4.27225 0.011021515 0.015988959578303513 0 +607 0 -6.331703 0.0016699883 0.0024112977743737099 0 +608 1 8.117621 0.9990056 0.0014353070238773355 1 +609 0 -4.0549765 0.013429949 0.019506602422452432 0 +610 1 8.266863 0.99913317 0.0012511141759918355 1 +611 1 6.1478367 0.9939311 0.0087822269655119054 1 +612 1 14.358189 0.99999684 4.5575474466196295E-06 1 +613 0 -4.861019 0.006438565 0.0093189182871853005 0 +614 0 -5.3180437 0.004236212 0.0061245436560297874 0 +615 0 -3.94002 0.014907227 0.021668495199742271 0 +616 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +617 0 ? ? ? 0 +618 0 -4.338106 0.010379728 0.015053041101426466 0 +619 0 -3.7689242 0.017406764 0.025333786303575221 0 +620 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +621 0 -0.16198826 0.32918024 0.57600290997485726 0 +622 0 -1.7370644 0.103187285 0.15712136267093918 0 +623 0 -6.331703 0.0016699883 0.0024112977743737099 0 +624 0 -3.4554825 0.023096563 0.033712130719424638 0 +625 0 -3.6625319 0.019164125 0.027916346227761391 0 +626 1 4.185855 0.96414673 0.052675368274755011 1 +627 0 -4.349147 0.010275813 0.014901558812792995 0 +628 0 -5.7625217 0.0028173493 0.0040703122701123903 0 +629 0 -4.6241584 0.007995273 0.01158109965570036 0 +630 0 -2.7987108 0.041490424 0.061135250300756741 0 +631 0 -3.7689242 0.017406764 0.025333786303575221 0 +632 0 -6.331703 0.0016699883 0.0024112977743737099 0 +633 1 3.9723816 0.9566957 0.06386801737041177 1 +634 0 -5.19334 0.004749254 0.0068680469780405835 0 +635 0 -4.5400214 0.008633794 0.01251001466156154 0 +636 1 9.007486 0.99956155 0.0006326909022333528 1 +637 0 -2.404181 0.058601033 0.087121824844992937 0 +638 0 -4.6241584 0.007995273 0.01158109965570036 0 +639 0 -3.3673954 0.02499933 0.036524884369080135 0 +640 0 -3.9203415 0.015175686 0.02206171472630955 0 +641 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +642 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +643 0 -6.331703 0.0016699883 0.0024112977743737099 0 +644 0 -5.7625217 0.0028173493 0.0040703122701123903 0 +645 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +646 0 -5.8342986 0.0026376308 0.0038103241926819645 0 +647 0 -5.627887 0.0031880299 0.0046067019330657986 0 +648 1 9.190195 0.99962944 0.00053470714742229219 1 +649 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +650 0 -3.9398255 0.014909858 0.021672348355527761 0 +651 0 -5.1132717 0.005110794 0.0073922236643752109 0 +652 0 -3.5992837 0.020290107 0.029573485680359592 0 +653 0 -4.338106 0.010379728 0.015053041101426466 0 +654 0 -4.621235 0.008016653 0.011612193952652979 0 +655 0 -4.9072876 0.0061716614 0.0089314146085223197 0 +656 0 -3.7689242 0.017406764 0.025333786303575221 0 +657 0 -0.46601295 0.2705457 0.45511048083845856 0 +658 1 8.120775 0.9990085 0.0014311753376832121 1 +659 0 -6.331703 0.0016699883 0.0024112977743737099 0 +660 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +661 0 -4.0520535 0.01346566 0.019558824244149512 0 +662 0 -5.5923867 0.0032936202 0.0047595318690019832 0 +663 0 -5.5923867 0.0032936202 0.0047595318690019832 0 +664 0 -4.203471 0.0117336875 0.017028231239157775 0 +665 0 -6.331703 0.0016699883 0.0024112977743737099 0 +666 0 -2.9374218 0.03669786 0.05393972532442147 0 +667 0 -4.621235 0.008016653 0.011612193952652979 0 +668 1 2.1150274 0.7997754 0.32233314938902585 1 +669 1 7.066202 0.9973858 0.0037764318699177366 1 +670 1 5.4009876 0.98799986 0.017417262659521413 1 +671 0 -4.0500174 0.013490589 0.019595281491330056 0 +672 0 -5.190417 0.0047619934 0.0068865141029330135 0 +673 0 -3.4082723 0.024098065 0.03519191086679558 0 +674 0 -6.0456505 0.0021721656 0.0031371809404331466 0 +675 0 -3.989233 0.014256162 0.020715307451098289 0 +676 0 -5.855971 0.0025856486 0.0037351333082833401 0 +677 0 -4.0549765 0.013429949 0.019506602422452432 0 +678 0 -6.331703 0.0016699883 0.0024112977743737099 0 +679 0 -5.7625217 0.0028173493 0.0040703122701123903 0 +680 1 14.850252 0.999998 2.9237080272005804E-06 1 +681 1 8.980978 0.9995507 0.00064834827339000788 1 +682 0 -3.3132315 0.02624428 0.038368198335972985 0 +683 0 -6.331703 0.0016699883 0.0024112977743737099 0 +684 0 -6.331703 0.0016699883 0.0024112977743737099 0 +685 0 -6.331703 0.0016699883 0.0024112977743737099 0 +686 0 -6.331703 0.0016699883 0.0024112977743737099 0 +687 0 -4.5680065 0.008415991 0.012193089858265733 0 +688 0 -4.6241584 0.007995273 0.01158109965570036 0 +689 0 -4.4540114 0.009338773 0.013536306372392479 0 +690 0 -5.627887 0.0031880299 0.0046067019330657986 0 +691 1 4.4992037 0.97288966 0.039651900791925578 1 +692 0 -5.19334 0.004749254 0.0068680469780405835 0 +693 0 -4.6390457 0.007887273 0.011424041506677902 0 +694 0 -4.998421 0.0056776917 0.0082145194870921444 0 +695 0 -5.7625217 0.0028173493 0.0040703122701123903 0 +696 1 6.3037586 0.9947386 0.0076106653397057331 1 +697 1 3.8583193 0.95212936 0.070770491793538823 1 +698 1 4.445225 0.97154737 0.041643762631074922 1 diff --git a/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Default-CV-breast-cancer.txt b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Default-CV-breast-cancer.txt new file mode 100644 index 0000000000..acf625e802 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Default-CV-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +5 1 7.057208 0.99996483 5.0735774115656486E-05 1 +6 0 -0.56973577 0.32680357 0.57090055823432817 0 +8 0 -3.265584 0.009919914 0.014382868280470263 0 +9 0 -2.8438797 0.018053178 0.026283197990078042 0 +10 0 -3.4260702 0.007889866 0.01142781187212402 0 +11 0 -3.2730165 0.009815386 0.014230562120318395 0 +18 1 4.1297426 0.9976289 0.0034247967532404057 1 +20 1 3.6110172 0.99501014 0.0072168704068302217 1 +21 1 3.7079868 0.9956574 0.0062787129039152982 1 +25 1 0.85096335 0.7895742 0.34085323485937402 1 +28 0 -3.2730165 0.009815386 0.014230562120318395 0 +31 0 -3.1337807 0.011967557 0.017369679358421687 0 +32 1 3.221815 0.9912949 0.012613756987139313 1 +35 0 -3.2730165 0.009815386 0.014230562120318395 0 +37 0 -2.1029623 0.050705004 0.07507161729782677 0 +40 0 ? ? ? 0 +41 1 0.4294026 0.67162436 0.57427353141509008 1 +44 1 4.518715 0.9986441 0.0019574580164138182 1 +45 0 -3.4117298 0.0080531 0.011665200864927582 0 +46 1 3.4790392 0.9939724 0.0087222723575821337 1 +48 0 -2.6620011 0.023330001 0.034056913922420239 0 +50 1 0.44066334 0.67518926 0.56663614622582026 1 +51 1 -0.21027565 0.44886696 1.1556401770975329 0 +52 1 2.103271 0.9579125 0.062034208068952469 1 +54 1 3.5159762 0.9942828 0.0082718705836708154 1 +56 1 3.9441147 0.99690485 0.0044722823151831939 1 +60 1 0.96663547 0.81590796 0.2935216879238205 1 +63 1 -0.135355 0.4756681 1.0719728110289044 0 +64 0 -3.4788275 0.007317102 0.010595157208169217 0 +66 0 -2.900283 0.016668824 0.024250712079095762 0 +68 1 4.283942 0.9981 0.0027437528959210881 1 +69 0 -3.2245307 0.010517443 0.015253819219210365 0 +70 0 -2.5950685 0.025629146 0.037457115844697367 0 +71 1 3.5624905 0.9946511 0.0077375738770138498 1 +72 0 -1.5871629 0.10090377 0.15345256258263343 0 +73 1 4.728806 0.9989976 0.0014468413771274271 1 +74 1 0.4975586 0.6928865 0.529309106252146 1 +76 0 -2.8515797 0.017857738 0.025996082517771347 0 +77 0 -2.3885822 0.034196626 0.050198591953806257 0 +79 0 -3.1797595 0.011209723 0.016263537352781413 0 +82 0 -2.620762 0.024721572 0.036113948744912555 0 +88 0 -2.900283 0.016668824 0.024250712079095762 0 +90 0 -3.3328135 0.009013158 0.013062193374478125 0 +91 0 -3.1149797 0.012291824 0.017843243228432587 0 +92 0 -2.900283 0.016668824 0.024250712079095762 0 +93 0 -3.4788275 0.007317102 0.010595157208169217 0 +95 0 -3.3328135 0.009013158 0.013062193374478125 0 +96 0 -3.3179266 0.009206588 0.013343819616380256 0 +97 0 -2.7835345 0.019659273 0.028644836222338311 0 +98 1 4.165916 0.9977489 0.0032512952389382424 1 +99 1 5.3154254 0.99956894 0.00062202334009018689 1 +100 1 2.4310963 0.9733226 0.03901008486188097 1 +102 0 -2.7532477 0.020517334 0.029908134186775407 0 +104 1 5.4333305 0.9996362 0.00052498655830247973 1 +105 1 -0.1526835 0.46945107 1.0909532990735458 0 +106 1 5.426038 0.99963236 0.00053049201666395714 1 +108 0 -3.2805295 0.009710832 0.014078235404188764 0 +109 1 3.3400679 0.9926474 0.010646734621373733 1 +111 1 2.1333807 0.9596256 0.059456447206265688 1 +112 1 4.0789456 0.9974495 0.0036842691473833646 1 +113 1 5.289747 0.99955267 0.00064550928688674445 1 +115 0 -2.383574 0.034435526 0.050555498183646824 0 +117 1 4.4668283 0.9985391 0.0021091884287959238 1 +120 0 -2.9493718 0.015549318 0.02260916150857321 0 +121 0 -2.4117963 0.03311001 0.048576343639436574 0 +122 1 5.8156424 0.99979013 0.00030280723272274124 1 +123 1 2.3205795 0.96886563 0.045631495325782319 1 +125 0 -3.4788275 0.007317102 0.010595157208169217 0 +128 1 2.6282647 0.97978115 0.029468557791152337 1 +129 0 -2.7184992 0.021547018 0.031425568220799491 0 +131 0 -3.1337807 0.011967557 0.017369679358421687 0 +132 1 5.7164726 0.99975795 0.00034925304341121586 1 +133 0 -3.039434 0.013684562 0.019878980259628184 0 +137 0 -3.2418132 0.010261675 0.014880951130273989 0 +138 0 -2.8588524 0.017675051 0.025727754571605383 0 +141 0 -3.4372978 0.0077643595 0.01124531595574753 0 +144 0 -3.2730165 0.009815386 0.014230562120318395 0 +145 0 ? ? ? 0 +147 0 -3.020754 0.0140522765 0.020416940217306098 0 +150 0 -3.4260702 0.007889866 0.01142781187212402 0 +151 1 2.2177572 0.9640774 0.052779098861998497 1 +152 1 5.1777344 0.99947447 0.00075838481234789161 1 +154 0 -3.6054602 0.006105267 0.0088350360454764378 0 +156 0 -3.0233922 0.01399976 0.020340097341932664 0 +161 0 -2.8170805 0.018749898 0.027307196469425622 0 +164 0 ? ? ? 0 +167 1 2.9291844 0.9867951 0.019177502987692791 1 +169 0 -3.6017787 0.006137507 0.0088818350941332908 0 +171 0 -3.3328135 0.009013158 0.013062193374478125 0 +173 1 7.947592 0.9999902 1.4102646298686894E-05 1 +174 1 3.2498922 0.9916369 0.01211617836068587 1 +176 0 -3.1337807 0.011967557 0.017369679358421687 0 +177 1 3.3192258 0.9924252 0.010969720954957895 1 +179 1 1.3783615 0.8890948 0.16959079187424589 1 +180 0 -3.4260702 0.007889866 0.01142781187212402 0 +181 0 -3.6054602 0.006105267 0.0088350360454764378 0 +183 1 5.419509 0.9996289 0.00053548135645133934 1 +187 1 5.8260784 0.99979323 0.00029833475210104223 1 +188 1 4.9982443 0.9993196 0.00098192506222582999 1 +189 0 -2.6896749 0.022439359 0.032741894658197351 0 +191 1 6.3130674 0.9998974 0.00014799866201853666 1 +192 0 -2.889139 0.01693381 0.024639538118855928 0 +196 0 2.9049587 0.9863329 6.1931483378075782 1 +198 0 -3.6054602 0.006105267 0.0088350360454764378 0 +199 0 -3.1226366 0.01215873 0.01764885214439868 0 +201 1 5.4355135 0.9996373 0.00052335212939580867 1 +202 0 -3.3328135 0.009013158 0.013062193374478125 0 +204 0 -3.3328135 0.009013158 0.013062193374478125 0 +205 1 7.059002 0.99996495 5.0563785426763426E-05 1 +206 1 3.804298 0.9962174 0.0054674425275320912 1 +207 0 -3.4260702 0.007889866 0.01142781187212402 0 +209 0 -2.88948 0.016925637 0.024627543492653111 0 +210 1 7.6166453 0.99998426 2.2701888528575377E-05 1 +211 1 5.4764957 0.9996581 0.00049333058043002202 1 +212 0 -3.3328135 0.009013158 0.013062193374478125 0 +216 0 -3.4788275 0.007317102 0.010595157208169217 0 +218 1 4.516878 0.99864054 0.0019626245103572994 1 +219 0 -2.360191 0.035572376 0.052255120544186871 0 +223 1 2.745268 0.9828614 0.024940108774080157 1 +226 1 4.9640293 0.99928534 0.0010314045874782159 1 +228 0 -3.4260702 0.007889866 0.01142781187212402 0 +233 1 3.1638048 0.99054396 0.013707089821426618 1 +237 1 3.3809423 0.99306464 0.010040464153854149 1 +239 1 2.5812135 0.9783949 0.031511173482079449 1 +240 0 -1.7316828 0.08353537 0.12584889741823041 0 +241 0 -2.8964548 0.01675939 0.024383591484353467 0 +242 0 -3.1337807 0.011967557 0.017369679358421687 0 +244 0 -3.3328135 0.009013158 0.013062193374478125 0 +246 1 6.6648927 0.9999382 8.9175760478377495E-05 1 +247 1 1.5379703 0.90980685 0.13636780191369618 1 +248 0 -2.3839004 0.034419905 0.050532159587373608 0 +249 0 ? ? ? 0 +250 0 -3.1919966 0.011016151 0.015981134131724528 0 +252 0 2.3607361 0.9705628 5.0862165055376689 1 +254 1 4.4944267 0.99859595 0.0020270350213282213 1 +257 0 -3.1226366 0.01215873 0.01764885214439868 0 +258 0 -3.017032 0.014126703 0.020525849645192915 0 +259 0 2.3594117 0.9705083 5.0835460984539447 1 +260 1 4.7412558 0.9990154 0.0014211904782310285 1 +262 1 5.535843 0.99968606 0.00045298741967439286 1 +267 1 1.7871778 0.93523324 0.096601895090605155 1 +268 1 3.395525 0.9932077 0.0098326586301602011 1 +269 0 -3.3328135 0.009013158 0.013062193374478125 0 +271 0 -2.7835345 0.019659273 0.028644836222338311 0 +272 1 1.7871778 0.93523324 0.096601895090605155 1 +275 0 ? ? ? 0 +276 0 -3.1226366 0.01215873 0.01764885214439868 0 +277 0 -3.4788275 0.007317102 0.010595157208169217 0 +278 0 -3.3328135 0.009013158 0.013062193374478125 0 +279 1 2.5009673 0.97581345 0.035322728606119444 1 +280 0 -3.017032 0.014126703 0.020525849645192915 0 +283 1 2.8390663 0.9849935 0.021813865188733895 1 +284 1 4.045544 0.9973242 0.0038654964837656733 1 +285 1 7.694996 0.999986 2.0208103005084558E-05 1 +288 1 0.65054774 0.7376656 0.43896114819175791 1 +290 0 -3.6054602 0.006105267 0.0088350360454764378 0 +291 0 -3.3328135 0.009013158 0.013062193374478125 0 +293 1 2.8659725 0.9855554 0.020991109422172988 1 +296 0 0.84631276 0.7884598 2.2409960920377765 1 +297 0 ? ? ? 0 +299 1 3.3291118 0.9925314 0.010815323080105525 1 +300 1 3.616962 0.99505246 0.0071555116925496928 1 +301 0 -3.3328135 0.009013158 0.013062193374478125 0 +303 0 -3.3328135 0.009013158 0.013062193374478125 0 +304 1 3.2374554 0.9914871 0.012334113503203046 1 +308 1 3.3689504 0.9929447 0.010214697519032298 1 +309 0 -1.5727416 0.102802716 0.15650284210453894 0 +311 0 -3.6054602 0.006105267 0.0088350360454764378 0 +312 1 1.5964963 0.9164854 0.12581615132950291 1 +314 0 -3.5903516 0.006238662 0.0090286797944395432 0 +316 1 2.323201 0.96897924 0.045462338889384266 1 +317 1 5.185748 0.9994805 0.00074969514792004902 1 +319 0 1.087976 0.84071136 2.6502846697574731 1 +321 0 ? ? ? 0 +323 1 3.347725 0.99272746 0.010530397548729942 1 +327 0 -3.4788275 0.007317102 0.010595157208169217 0 +328 1 2.3924484 0.97183937 0.04121021896301344 1 +329 1 3.653625 0.9953056 0.0067885373582329865 1 +331 0 -2.764847 0.020184463 0.02941792601227879 0 +332 0 -2.330146 0.03708628 0.054521558197439139 0 +333 1 3.0650225 0.9891148 0.0157900895649082 1 +336 1 3.4496598 0.99371374 0.0090977867244981265 1 +338 0 -3.5903516 0.006238662 0.0090286797944395432 0 +343 0 -3.6054602 0.006105267 0.0088350360454764378 0 +344 1 5.3123817 0.99956703 0.00062477625180167376 1 +346 0 -2.2415557 0.04191894 0.061780373479293986 0 +347 0 -3.4701545 0.0074083433 0.010727766812827805 0 +348 1 0.009417534 0.52771896 0.92215827358141711 1 +349 1 1.9216323 0.94601464 0.080065580657279145 1 +350 0 -2.7972674 0.019281887 0.028089570798842188 0 +352 0 0.48690295 0.689613 1.6878598894644639 1 +353 1 5.68491 0.9997467 0.00036550943041529892 1 +354 0 -3.4788275 0.007317102 0.010595157208169217 0 +355 0 -2.8206 0.018656915 0.027170493184520649 0 +358 1 3.4298716 0.99353325 0.0093598394390381444 1 +360 1 8.039464 0.9999915 1.2296811942483466E-05 1 +361 1 4.0018487 0.997151 0.0041160794808710478 1 +366 1 7.6356616 0.99998474 2.2013947263955502E-05 1 +368 0 -3.1859484 0.011111408 0.01612009782603635 0 +370 0 -2.4288907 0.032331247 0.047414818068910486 0 +371 0 -3.1859484 0.011111408 0.01612009782603635 0 +373 0 -2.7758775 0.019872827 0.028959142296125673 0 +376 0 -3.4788275 0.007317102 0.010595157208169217 0 +377 0 -3.5903516 0.006238662 0.0090286797944395432 0 +378 0 -2.5109992 0.02883105 0.04220579778763664 0 +379 0 -1.6979226 0.08733187 0.13183773892539827 0 +381 1 4.533961 0.9986735 0.0019150073584088983 1 +383 0 -3.4372978 0.0077643595 0.01124531595574753 0 +384 0 -3.4372978 0.0077643595 0.01124531595574753 0 +387 0 -1.8698577 0.06951584 0.10394650275159649 0 +388 0 -3.164734 0.01145201 0.016617090089496449 0 +389 0 -2.384867 0.034373697 0.050463120073904313 0 +391 1 5.278062 0.9995451 0.00065643511405119592 1 +392 0 -3.1226366 0.01215873 0.01764885214439868 0 +395 0 -3.1226366 0.01215873 0.01764885214439868 0 +396 0 -3.017032 0.014126703 0.020525849645192915 0 +398 0 -2.8773105 0.017219596 0.025059002727364624 0 +399 0 -2.983367 0.014817738 0.021537442200535013 0 +404 0 -3.035563 0.013759975 0.019989291821870307 0 +406 0 -2.6738753 0.022943696 0.033486393337738882 0 +409 0 -2.9756012 0.014981812 0.02177773104095828 0 +413 0 -2.636499 0.02418125 0.035314891213896034 0 +414 1 3.5129192 0.9942577 0.0083082815579947292 1 +415 0 -0.44547534 0.3673017 0.66041038111257111 0 +416 1 5.0948877 0.99940795 0.00085440478690671634 1 +418 0 -1.7674733 0.0796749 0.11978451859048206 0 +419 0 -2.7845545 0.019630998 0.028603226661865832 0 +422 0 -1.9319801 0.06394965 0.09534196376128598 0 +423 0 -2.5950685 0.025629146 0.037457115844697367 0 +428 0 -3.4788275 0.007317102 0.010595157208169217 0 +429 0 -3.2730165 0.009815386 0.014230562120318395 0 +430 0 -2.9450116 0.015645685 0.022750392173165056 0 +434 0 3.337569 0.9926211 7.0823831717344676 1 +436 1 2.4818127 0.9751541 0.036297871506884037 1 +439 0 -3.0091302 0.014285991 0.020758965120568837 0 +440 1 4.7098246 0.99896985 0.0014869540997041347 1 +441 0 -1.2440159 0.15534541 0.24356660358453483 0 +442 0 -2.68195 0.022684572 0.033103828921237276 0 +449 1 5.942175 0.99982506 0.0002524066187454977 1 +450 0 -2.8853104 0.017025797 0.024774539211150356 0 +451 0 -3.0091302 0.014285991 0.020758965120568837 0 +452 0 -3.001718 0.01443702 0.020980028506503847 0 +453 1 4.8444138 0.9991512 0.0012251224992695688 1 +454 0 -2.9710684 0.015078406 0.02191921319416805 0 +455 1 -0.41451907 0.377717 1.4046224225171875 0 +456 1 5.1211615 0.9994299 0.00082274158011057654 1 +457 1 4.6062374 0.9988044 0.0017259319729422537 1 +464 0 -3.1254716 0.012109812 0.017577411919819597 0 +465 1 5.048981 0.99936754 0.00091274262341741255 1 +466 1 4.630231 0.998845 0.0016673029728197354 1 +467 1 4.1375737 0.9976554 0.003386526369283925 1 +474 0 -3.0091302 0.014285991 0.020758965120568837 0 +480 0 -3.1006896 0.012544082 0.018211749391609001 0 +482 1 7.2396455 0.999973 3.8954596327452357E-05 1 +483 1 5.804902 0.99978685 0.00030753775616068136 1 +484 0 -2.866447 0.017486239 0.025450480899592006 0 +487 1 6.835632 0.99995166 6.9740650547897242E-05 1 +489 1 -0.97964525 0.21203105 2.2376525348702296 0 +492 0 -3.1118336 0.012346929 0.017923733802516856 0 +493 1 6.144757 0.9998693 0.00018859130260575588 1 +495 0 -3.2285826 0.010456917 0.01516557361150184 0 +497 0 -2.979243 0.014904646 0.021664715707197923 0 +501 0 -3.005888 0.014351862 0.020855376795789662 0 +502 0 -2.8926263 0.016850445 0.024517201983634777 0 +504 0 -3.6054602 0.006105267 0.0088350360454764378 0 +507 0 -2.6673017 0.02315678 0.033801062847352965 0 +510 0 -3.6054602 0.006105267 0.0088350360454764378 0 +513 0 -3.2285826 0.010456917 0.01516557361150184 0 +514 1 5.730318 0.9997627 0.00034237208821625534 1 +517 0 -3.5903516 0.006238662 0.0090286797944395432 0 +519 1 3.836201 0.9963866 0.0052224933253354131 1 +520 0 -3.4750462 0.0073567443 0.010652771541403745 0 +521 0 -3.0694122 0.0131141795 0.019044915681778928 0 +522 1 1.8001244 0.9363529 0.094875715137687208 1 +523 1 4.2548246 0.9980188 0.0028611008075221564 1 +527 0 -2.900283 0.016668824 0.024250712079095762 0 +528 0 -2.4706626 0.030502046 0.044690240469598054 0 +529 0 -3.1118336 0.012346929 0.017923733802516856 0 +531 0 -2.6738753 0.022943696 0.033486393337738882 0 +532 0 -3.4260702 0.007889866 0.01142781187212402 0 +533 0 -3.1226366 0.01215873 0.01764885214439868 0 +534 0 -3.2730165 0.009815386 0.014230562120318395 0 +535 0 -2.7646904 0.020188924 0.029424494522034134 0 +538 0 -3.005888 0.014351862 0.020855376795789662 0 +539 0 -2.7723904 0.019970834 0.029103410303146709 0 +540 0 -2.6079202 0.025171202 0.036779223201022067 0 +541 0 -3.2418132 0.010261675 0.014880951130273989 0 +544 0 -2.7833471 0.019664474 0.028652489442122082 0 +546 1 6.903044 0.99995613 6.3291003784773738E-05 1 +547 0 -3.5305548 0.006795633 0.0098374901745714467 0 +548 0 -3.3341873 0.0089955125 0.013036504600974122 0 +549 1 3.0349069 0.98863804 0.016485671543673565 1 +557 0 -2.7972674 0.019281887 0.028089570798842188 0 +558 0 -3.2730165 0.009815386 0.014230562120318395 0 +559 0 -2.889139 0.01693381 0.024639538118855928 0 +560 0 -2.7835345 0.019659273 0.028644836222338311 0 +561 0 -2.7835345 0.019659273 0.028644836222338311 0 +563 0 -3.1226366 0.01215873 0.01764885214439868 0 +565 1 6.7825165 0.9999478 7.533036773246141E-05 1 +566 0 -2.8699965 0.017398672 0.025321906235678267 0 +569 1 5.348246 0.9995888 0.00059337616434588982 1 +577 0 -3.4788275 0.007317102 0.010595157208169217 0 +578 0 -3.4788275 0.007317102 0.010595157208169217 0 +581 1 5.0124693 0.99933344 0.00096196162989513302 1 +582 1 5.0096254 0.9993307 0.0009659198746940407 1 +584 0 -2.2400014 0.042008888 0.061915823789131894 0 +586 1 7.861235 0.999989 1.5908482915272255E-05 1 +590 1 2.2271297 0.96454173 0.052084433521439384 1 +593 0 -2.866447 0.017486239 0.025450480899592006 0 +594 1 3.2574682 0.9917268 0.011985329025336479 1 +600 0 -3.1226366 0.01215873 0.01764885214439868 0 +602 0 -3.005888 0.014351862 0.020855376795789662 0 +604 1 2.9559016 0.987287 0.018458586152655318 1 +606 0 -2.998231 0.014508615 0.021084835723885111 0 +607 0 -3.6054602 0.006105267 0.0088350360454764378 0 +609 0 -3.0091302 0.014285991 0.020758965120568837 0 +612 1 8.095181 0.99999213 1.135089961038231E-05 1 +613 0 -2.8651934 0.017517269 0.025496044896015015 0 +614 0 -3.3662734 0.008592977 0.012450616450921022 0 +617 0 ? ? ? 0 +618 0 -3.005888 0.014351862 0.020855376795789662 0 +619 0 -2.889139 0.01693381 0.024639538118855928 0 +621 0 -0.85922766 0.24243243 0.40055352247503417 0 +622 0 -2.2535048 0.04123357 0.060748698239818967 0 +624 0 -2.8818233 0.017110007 0.024898138409667345 0 +627 0 -2.2639332 0.04064419 0.059862109278937758 0 +629 0 -3.1254716 0.012109812 0.017577411919819597 0 +633 1 2.2230194 0.9643388 0.052388029791135139 1 +634 0 -3.2418132 0.010261675 0.014880951130273989 0 +638 0 -3.1254716 0.012109812 0.017577411919819597 0 +639 0 -2.7972674 0.019281887 0.028089570798842188 0 +641 0 -3.1226366 0.01215873 0.01764885214439868 0 +642 0 -3.1226366 0.01215873 0.01764885214439868 0 +644 0 -3.4372978 0.0077643595 0.01124531595574753 0 +645 0 -3.1226366 0.01215873 0.01764885214439868 0 +649 0 -3.1226366 0.01215873 0.01764885214439868 0 +652 0 -2.8059366 0.01904732 0.027744550507937341 0 +653 0 -3.005888 0.014351862 0.020855376795789662 0 +654 0 -3.017032 0.014126703 0.020525849645192915 0 +656 0 -2.889139 0.01693381 0.024639538118855928 0 +657 0 -0.18228102 0.4588551 0.88591312504644681 0 +660 0 -3.4788275 0.007317102 0.010595157208169217 0 +661 0 -2.900283 0.016668824 0.024250712079095762 0 +665 0 -3.6054602 0.006105267 0.0088350360454764378 0 +668 1 1.1573684 0.8536372 0.22830501603144154 1 +670 1 4.2615523 0.9980379 0.002833529221533669 1 +678 0 -3.6054602 0.006105267 0.0088350360454764378 0 +679 0 -3.4372978 0.0077643595 0.01124531595574753 0 +680 1 8.287129 0.99999404 8.5991581706542968E-06 1 +681 1 5.9939256 0.99983764 0.00023425938852449468 1 +682 0 -2.7003317 0.022105314 0.032248991603681956 0 +683 0 -3.6054602 0.006105267 0.0088350360454764378 0 +685 0 -3.6054602 0.006105267 0.0088350360454764378 0 +688 0 -3.1254716 0.012109812 0.017577411919819597 0 +689 0 -3.2345 0.010369146 0.015037614915942544 0 +691 1 3.3198962 0.9924324 0.010959236625725764 1 +692 0 -3.2418132 0.010261675 0.014880951130273989 0 +693 0 -3.1039493 0.012486094 0.018127030314179445 0 +694 0 -2.8741245 0.017297378 0.025173189409213471 0 +696 1 4.167007 0.9977525 0.0032461241280301052 1 +697 1 2.6165218 0.97944355 0.02996574922798281 1 +698 1 3.2709067 0.991884 0.011756696347368025 1 +0 0 -2.1858048 0.019522201 0.028443131501617723 0 +1 0 2.4643466 0.87950635 3.0529709710484978 1 +2 0 -2.3502047 0.015902877 0.023127388909652609 0 +3 0 2.4278917 0.87451434 2.9944056022897438 1 +4 0 -2.082891 0.022186847 0.032369283673121328 0 +7 0 -2.638155 0.011087011 0.016084505454470342 0 +12 1 0.13348699 0.27454063 1.8649084059985652 1 +13 0 -2.4949718 0.013268124 0.019269979978932417 0 +14 1 5.299982 0.99627197 0.0053884638761153361 1 +15 1 0.7744734 0.46061882 1.1183547250072077 1 +16 0 -2.4918776 0.013319657 0.019345327864414775 0 +17 0 -2.4005723 0.014932416 0.021705386349413373 0 +19 0 -1.9598849 0.025840316 0.037769816954696993 0 +22 0 -2.7270536 0.009915355 0.014376224049015385 0 +23 1 ? ? ? 0 +24 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +26 0 -2.5670438 0.012121931 0.017595109452200379 0 +27 0 -2.280314 0.017352993 0.025254839289623091 0 +29 0 -2.9168253 0.0078088576 0.011310016984425502 0 +30 0 -2.743445 0.009713114 0.014081560900358588 0 +33 0 -2.697853 0.01028602 0.014916437796865623 0 +34 0 -2.5725198 0.012038953 0.017473933576498117 0 +36 1 5.370757 0.99659127 0.0049261586606575156 1 +38 1 3.8173842 0.97600687 0.035036798748799364 1 +39 1 1.7423136 0.74478805 0.42509816834409625 1 +42 1 5.30029 0.9962734 0.0053863923631196743 1 +43 1 0.4313023 0.35581556 1.490798494614201 1 +47 0 -3.1699133 0.005674961 0.0082105575200780885 0 +49 1 4.063138 0.98232436 0.025728616279702864 1 +53 1 3.6499045 0.9704895 0.043215487885653277 1 +55 1 3.0962358 0.9421367 0.085991683453461656 1 +57 1 0.31318235 0.3222287 1.6338430986952079 1 +58 1 1.4862127 0.6782709 0.56006654718119175 1 +59 1 0.8486433 0.48408565 1.0466657688355494 1 +61 0 -2.9074705 0.007901425 0.011444620183650155 0 +62 1 4.689612 0.9919433 0.01167043745866826 1 +65 1 1.4220567 0.6602432 0.59893052828159399 1 +67 1 3.0339043 0.93766826 0.092850490546770872 1 +75 0 -2.2670527 0.017642453 0.025679879972587048 0 +78 0 -2.0084548 0.024332521 0.035538554511704831 0 +80 0 -2.03422 0.023567848 0.03440829372503814 0 +81 0 -2.310827 0.016704572 0.024303160568656917 0 +83 0 -1.8550333 0.029411687 0.043068605771028841 0 +84 1 5.2740946 0.9961479 0.0055681789505025291 1 +85 1 3.909911 0.97860944 0.031194891325044546 1 +86 1 1.9654632 0.79483765 0.33126787626487048 1 +87 1 4.4889193 0.98962927 0.015039925710551912 1 +89 0 -2.795878 0.009093219 0.013178752422657699 0 +94 0 -2.8278563 0.008734566 0.012656671370376122 0 +101 1 0.26796246 0.30981943 1.6905004702937003 1 +103 1 0.23171926 0.3000665 1.7366458278668633 1 +107 1 4.468525 0.9893601 0.015432385453245953 1 +110 0 -1.4909216 0.04590447 0.067794367893782667 0 +114 0 -1.425511 0.04968218 0.073518014347900285 0 +116 0 -0.30933404 0.1774163 0.28176560155499253 0 +118 0 -2.611174 0.011468971 0.016641843684379737 0 +119 0 -2.1002054 0.02171489 0.03167311202362464 0 +124 1 4.1922474 0.9849567 0.021867818470544111 1 +126 1 4.3657174 0.9878945 0.017571150102842548 1 +127 0 -2.6042523 0.011569039 0.016787893374491106 0 +130 0 -1.9203931 0.027132992 0.039685493973052155 0 +134 0 -2.7086492 0.010147406 0.014714395385468908 0 +135 0 -1.4709824 0.0470261 0.069491394938931556 0 +136 0 -2.4918776 0.013319657 0.019345327864414775 0 +139 0 ? ? ? 0 +140 0 -2.8614268 0.008373138 0.012130741394297474 0 +142 1 2.5781393 0.89399797 0.16165654384957087 1 +143 0 -2.1831431 0.019586993 0.028538471086822469 0 +146 1 1.1117847 0.5671983 0.81807494605491005 1 +148 0 -0.7938956 0.104406536 0.15908409477729052 0 +149 1 6.315873 0.99897087 0.0014854907404379277 1 +153 0 -2.10214 0.021662772 0.031596253704097708 0 +155 1 2.5817873 0.8944361 0.1609496436064439 1 +157 0 -2.8278563 0.008734566 0.012656671370376122 0 +158 0 ? ? ? 0 +159 1 7.2913313 0.9997015 0.00043070884426719654 1 +160 1 5.656353 0.9976256 0.0034296237206113307 1 +162 0 -2.6042523 0.011569039 0.016787893374491106 0 +163 0 -2.2131763 0.018867977 0.027480813524908082 0 +165 0 -1.9289669 0.02684711 0.039261612108107226 0 +166 1 5.371228 0.9965933 0.0049232249583664065 1 +168 0 -2.6042523 0.011569039 0.016787893374491106 0 +170 0 -2.8614268 0.008373138 0.012130741394297474 0 +172 0 -3.1699133 0.005674961 0.0082105575200780885 0 +175 1 5.180047 0.99566144 0.0062728400019319355 1 +178 0 -2.4005723 0.014932416 0.021705386349413373 0 +182 0 -1.9598849 0.025840316 0.037769816954696993 0 +184 1 3.955361 0.9797846 0.029463467380837841 1 +185 0 -2.8716645 0.008265896 0.011974726218235726 0 +186 1 3.985546 0.98052996 0.028366376544727477 1 +190 1 7.4657273 0.9997608 0.00034512446635598279 1 +193 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +194 0 -2.6042523 0.011569039 0.016787893374491106 0 +195 0 -2.4005723 0.014932416 0.021705386349413373 0 +197 0 -1.8385303 0.030015778 0.043966814323580343 0 +200 1 6.0277863 0.99851704 0.0021410521205442053 1 +203 0 -2.1858048 0.019522201 0.028443131501617723 0 +208 0 -3.0978432 0.0062153796 0.0089948796304217538 0 +213 1 8.651913 0.99994695 7.6534309653049576E-05 1 +214 1 8.021142 0.9998818 0.00017053087656366316 1 +215 1 4.766493 0.9926871 0.010589041348733017 1 +217 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +220 0 -2.9111686 0.0078647025 0.011391220512492049 0 +221 1 5.885932 0.9982249 0.0025631832715805789 1 +222 1 -1.0910792 0.07401945 3.7559518330884303 0 +224 1 5.8083706 0.99804157 0.0028281872876838774 1 +225 0 -3.1699133 0.005674961 0.0082105575200780885 0 +227 1 4.6679697 0.9917207 0.011994260047261323 1 +229 1 7.4762716 0.99976397 0.00034056584291574553 1 +230 1 3.553524 0.9667756 0.048747038591416381 1 +231 1 5.22736 0.9959133 0.0059079033199315846 1 +232 0 0.6768029 0.4299971 0.81095880149551824 1 +234 0 -1.2008077 0.06501901 0.0969910647234625 0 +235 0 ? ? ? 0 +236 1 6.9204216 0.9995221 0.00068964325415644641 1 +238 1 7.6631403 0.9998138 0.0002686619145078823 1 +243 0 -1.7283036 0.034369797 0.050457292720221654 0 +245 0 -1.7236626 0.0345659 0.050750309203949198 0 +251 1 4.627567 0.99128866 0.01262286539422033 1 +253 1 5.30029 0.9962734 0.0053863923631196743 1 +255 1 3.4353652 0.96160424 0.05648484121740667 1 +256 0 -2.8614268 0.008373138 0.012130741394297474 0 +261 1 6.7552137 0.9994106 0.00085053289008703693 1 +263 1 5.02655 0.99473274 0.007619137087775783 1 +264 1 3.0155752 0.93629414 0.094966268717484412 1 +265 0 -1.4002796 0.051216733 0.075849528101537778 0 +266 1 5.257723 0.9960673 0.00568489352489388 1 +270 1 4.0899096 0.982905 0.024876154420347706 1 +273 1 0.335526 0.3284555 1.606230137088891 1 +274 0 -2.4146152 0.014672402 0.021324629190085015 0 +281 0 -2.697853 0.01028602 0.014916437796865623 0 +282 1 3.025592 0.93704855 0.093804289898579224 1 +286 1 8.930486 0.99996275 5.3745579490272704E-05 1 +287 0 -2.7086492 0.010147406 0.014714395385468908 0 +289 1 5.334999 0.9964335 0.0051545743258560924 1 +292 1 ? ? ? 0 +294 0 ? ? ? 0 +295 1 4.5509434 0.9904071 0.013906424460647474 1 +298 0 -1.5676327 0.04182242 0.061635035893032077 0 +302 1 8.872599 0.9999599 5.7873322786187624E-05 1 +305 1 5.757324 0.9979107 0.003017407196179228 1 +306 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +307 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +310 0 -2.922761 0.007750682 0.011225429363134413 0 +313 0 -3.299302 0.0048193554 0.0069696683697878879 0 +315 0 ? ? ? 0 +318 0 -2.8691468 0.008292142 0.012012908069216954 0 +320 1 3.933642 0.9792311 0.030278688471899282 1 +322 0 -2.6042523 0.011569039 0.016787893374491106 0 +324 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +325 0 -2.0716977 0.022497274 0.032827369134600271 0 +326 1 2.3247476 0.85942423 0.218557637107754 1 +330 1 2.9209402 0.92874014 0.10665309970829556 1 +334 1 4.1128078 0.9833867 0.024169258370237343 1 +335 0 -3.299302 0.0048193554 0.0069696683697878879 0 +337 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +339 1 3.6997185 0.972248 0.040603706580857601 1 +340 1 3.8200495 0.97608596 0.034919887823900035 1 +341 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +342 0 -3.111528 0.006108972 0.0088404137863029985 0 +345 0 -3.299302 0.0048193554 0.0069696683697878879 0 +351 0 -2.8278563 0.008734566 0.012656671370376122 0 +356 1 -0.28485298 0.18199818 2.4580040844636111 0 +357 1 6.906701 0.9995137 0.00070177387936437943 1 +359 1 3.2611704 0.95255077 0.070132107772800226 1 +362 0 -1.7083551 0.035220392 0.051728680827174772 0 +363 0 -0.7258245 0.11276888 0.1726181270358976 0 +364 0 -2.8278563 0.008734566 0.012656671370376122 0 +365 0 -2.977202 0.0072367736 0.010478418271730705 0 +367 1 6.7865734 0.9994336 0.0008174070864974929 1 +369 0 -2.8500295 0.008494151 0.012306811539543959 0 +372 0 -2.3130798 0.016657654 0.024234323683505889 0 +374 0 -2.5725198 0.012038953 0.017473933576498117 0 +375 0 -3.299302 0.0048193554 0.0069696683697878879 0 +380 0 -3.299302 0.0048193554 0.0069696683697878879 0 +382 0 -1.9081125 0.027547633 0.040300510056613802 0 +385 0 -1.8123745 0.030997915 0.045428325287406507 0 +386 1 3.5814629 0.96789634 0.047075545827350976 1 +390 0 -3.0397084 0.0066883285 0.0096816314790603344 0 +393 0 -3.3468409 0.0045383475 0.0065623538118057389 0 +394 0 -2.53023 0.012694622 0.018431709676809287 0 +397 0 -2.6262507 0.0112539595 0.016328082658637364 0 +400 1 5.6133165 0.99749255 0.0036220259996406776 1 +401 0 -2.8614268 0.008373138 0.012130741394297474 0 +402 0 -1.6083864 0.039797306 0.058589111268187659 0 +403 0 -2.0118618 0.024230035 0.035387018309250291 0 +405 0 -3.1699133 0.005674961 0.0082105575200780885 0 +407 0 -3.1699133 0.005674961 0.0082105575200780885 0 +408 0 -1.7974956 0.031570416 0.046280942730234045 0 +410 0 -3.1699133 0.005674961 0.0082105575200780885 0 +411 0 ? ? ? 0 +412 1 5.1801887 0.9956622 0.0062717172439807865 1 +417 0 -3.1699133 0.005674961 0.0082105575200780885 0 +420 0 -1.3703549 0.053094853 0.078708178587860597 0 +421 1 6.7476807 0.99940497 0.00085870690667158795 1 +424 0 -2.8614268 0.008373138 0.012130741394297474 0 +425 1 8.440741 0.9999306 0.00010009737521057061 1 +426 0 -1.1437708 0.06956262 0.10401903933106471 0 +427 1 3.0290432 0.9373065 0.093407171483572096 1 +431 0 -1.6003871 0.040187247 0.059175112811610614 0 +432 0 -2.0669918 0.022629047 0.033021865599754556 0 +433 0 -2.1939526 0.019325167 0.028153240295868898 0 +435 1 4.9303565 0.9940526 0.00860591734976527 1 +437 0 -2.6262507 0.0112539595 0.016328082658637364 0 +438 0 -1.9525014 0.026077354 0.038120904854391166 0 +443 0 -3.0108025 0.0069366256 0.01004230558528737 0 +444 0 -1.733223 0.0341631 0.050148510040325285 0 +445 0 -3.111528 0.006108972 0.0088404137863029985 0 +446 0 -3.299302 0.0048193554 0.0069696683697878879 0 +447 0 -2.391075 0.015110835 0.02196671623363939 0 +448 0 -3.3468409 0.0045383475 0.0065623538118057389 0 +458 0 -2.212277 0.018889127 0.027511914143950728 0 +459 0 -2.0334792 0.023589501 0.034440287123152627 0 +460 0 -2.1183214 0.021231595 0.030960564068005136 0 +461 0 -1.7272935 0.034412384 0.050520921795525156 0 +462 0 -1.9270117 0.026912043 0.039357879214519199 0 +463 0 -2.435658 0.014291106 0.020766451216628158 0 +468 0 -2.6262507 0.0112539595 0.016328082658637364 0 +469 0 -2.9698434 0.007304208 0.010576417869130071 0 +470 0 -2.743445 0.009713114 0.014081560900358588 0 +471 0 -1.9270117 0.026912043 0.039357879214519199 0 +472 0 -2.1959991 0.019275982 0.028080884826884558 0 +473 0 -2.6262507 0.0112539595 0.016328082658637364 0 +475 0 -2.8614268 0.008373138 0.012130741394297474 0 +476 0 -2.4474528 0.014081672 0.020459953791340746 0 +477 0 -2.6262507 0.0112539595 0.016328082658637364 0 +478 0 -2.2293873 0.018490694 0.026926149164253149 0 +479 1 5.2473936 0.9960156 0.0057597443048448315 1 +481 0 -1.4591035 0.04770666 0.070522051664368798 0 +485 0 -2.51326 0.012967533 0.018830554697715829 0 +486 0 -2.743445 0.009713114 0.014081560900358588 0 +488 1 1.20174 0.5949928 0.74905584451233209 1 +490 0 -3.299302 0.0048193554 0.0069696683697878879 0 +491 1 4.4116144 0.98857194 0.01658213512685398 1 +494 0 0.19906044 0.29143023 0.49701818839808581 1 +496 0 -3.3468409 0.0045383475 0.0065623538118057389 0 +498 0 -2.4918776 0.013319657 0.019345327864414775 0 +499 0 -2.4918776 0.013319657 0.019345327864414775 0 +500 0 -1.9598849 0.025840316 0.037769816954696993 0 +503 0 -2.4005723 0.014932416 0.021705386349413373 0 +505 0 -2.5028172 0.013138342 0.019080238066149724 0 +506 1 6.2301693 0.99885267 0.0016561973072822019 1 +508 0 -2.391075 0.015110835 0.02196671623363939 0 +509 0 -3.111528 0.006108972 0.0088404137863029985 0 +511 0 -2.280314 0.017352993 0.025254839289623091 0 +512 0 -2.391075 0.015110835 0.02196671623363939 0 +515 1 5.601015 0.99745315 0.0036790102734250547 1 +516 0 -3.3468409 0.0045383475 0.0065623538118057389 0 +518 0 -2.5321863 0.012663529 0.018386275703658014 0 +524 0 -2.7270536 0.009915355 0.014376224049015385 0 +525 0 -2.688909 0.010402273 0.015085908849059596 0 +526 0 -2.6262507 0.0112539595 0.016328082658637364 0 +530 1 4.0611787 0.9822811 0.025792170722406128 1 +536 0 -2.1858048 0.019522201 0.028443131501617723 0 +537 0 -1.9876449 0.02496774 0.036478141155914565 0 +542 0 -2.0680609 0.022599045 0.032977580948398179 0 +543 0 -2.4918776 0.013319657 0.019345327864414775 0 +545 0 -2.280314 0.017352993 0.025254839289623091 0 +550 0 -2.7270536 0.009915355 0.014376224049015385 0 +551 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +552 0 -1.7737393 0.032505758 0.047675019900819467 0 +553 0 -0.48280287 0.14751787 0.23025851084901836 0 +554 0 -2.8614268 0.008373138 0.012130741394297474 0 +555 0 -0.729087 0.11235509 0.17194543409253302 0 +556 0 -1.8535136 0.029466819 0.043150557662289729 0 +562 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +564 0 -2.168657 0.019943336 0.029062930891538222 0 +567 0 -1.9768028 0.025305066 0.036977349302418229 0 +568 1 3.1277177 0.94427764 0.082716980050333053 1 +570 1 4.791981 0.9929183 0.010253062861046026 1 +571 1 6.337371 0.9989985 0.001445550213597871 1 +572 0 -2.7270536 0.009915355 0.014376224049015385 0 +573 0 -3.1699133 0.005674961 0.0082105575200780885 0 +574 1 3.537343 0.9661093 0.049741714558953425 1 +575 0 -1.9876449 0.02496774 0.036478141155914565 0 +576 0 -2.280314 0.017352993 0.025254839289623091 0 +579 0 -3.0562034 0.0065506157 0.0094816299915327536 0 +580 0 -2.0864134 0.022090029 0.032226441907387202 0 +583 0 -2.8614268 0.008373138 0.012130741394297474 0 +585 0 -3.299302 0.0048193554 0.0069696683697878879 0 +587 0 -2.0669918 0.022629047 0.033021865599754556 0 +588 1 3.0994945 0.94236183 0.08564698778891873 1 +589 0 -2.391075 0.015110835 0.02196671623363939 0 +591 1 3.9345682 0.9792551 0.030243387214391203 1 +592 1 3.106784 0.9428625 0.084880684135415621 1 +595 0 -2.280314 0.017352993 0.025254839289623091 0 +596 0 -2.3130798 0.016657654 0.024234323683505889 0 +597 0 -1.7976105 0.031565957 0.046274299798906078 0 +598 0 -2.7270536 0.009915355 0.014376224049015385 0 +599 0 -1.7464846 0.03361184 0.049325317974486131 0 +601 0 -3.2164955 0.005350793 0.0077402890679609355 0 +603 1 2.4305253 0.87488085 0.19284154464442543 1 +605 1 5.5752525 0.9973687 0.0038011762787992279 1 +608 1 6.160969 0.99874747 0.0018081543347426264 1 +610 1 4.2562766 0.9861149 0.020172310679415974 1 +611 1 4.19181 0.98494846 0.021879866566343848 1 +615 0 -2.1282535 0.020971112 0.030576665061562713 0 +616 0 -2.7270536 0.009915355 0.014376224049015385 0 +620 0 -2.7270536 0.009915355 0.014376224049015385 0 +623 0 -3.299302 0.0048193554 0.0069696683697878879 0 +625 0 -1.6549138 0.0376002 0.055291753119665971 0 +626 1 3.2866921 0.9539941 0.067947758081130266 1 +628 0 -3.111528 0.006108972 0.0088404137863029985 0 +630 0 -1.7293918 0.034323968 0.050388824468901965 0 +631 0 -2.280314 0.017352993 0.025254839289623091 0 +632 0 -3.299302 0.0048193554 0.0069696683697878879 0 +635 0 -2.3237798 0.016436575 0.023910007293990707 0 +636 1 5.4205704 0.9967995 0.0046247087970721114 1 +637 0 -1.2474248 0.06151211 0.091589964164117119 0 +640 0 -2.3239968 0.016432121 0.023903474771901047 0 +643 0 -3.299302 0.0048193554 0.0069696683697878879 0 +646 0 -2.8692653 0.0082909055 0.01201110883078579 0 +647 0 -3.05244 0.0065817856 0.0095268959776332787 0 +648 1 6.285733 0.99893075 0.0015434236851713337 1 +650 0 -1.821827 0.030639436 0.044894703756425194 0 +651 0 -2.7272372 0.009913066 0.014372889731158276 0 +655 0 -2.7270536 0.009915355 0.014376224049015385 0 +658 1 5.176528 0.99564207 0.006300909234690056 1 +659 0 -3.299302 0.0048193554 0.0069696683697878879 0 +662 0 -2.790698 0.009152672 0.013265314439549255 0 +663 0 -2.790698 0.009152672 0.013265314439549255 0 +664 0 -2.5114403 0.012997138 0.018873827293542356 0 +666 0 -1.9454998 0.026304087 0.03845680802952664 0 +667 0 -2.6042523 0.011569039 0.016787893374491106 0 +669 1 5.3159976 0.9963468 0.0052801449977246768 1 +671 0 -2.322586 0.016461097 0.023945976311179954 0 +672 0 -2.8278563 0.008734566 0.012656671370376122 0 +673 0 -1.8290145 0.03036955 0.0444930893633559 0 +674 0 -3.1699133 0.005674961 0.0082105575200780885 0 +675 0 -2.0441403 0.02327972 0.033982644015074934 0 +676 0 -2.9698434 0.007304208 0.010576417869130071 0 +677 0 -2.391075 0.015110835 0.02196671623363939 0 +684 0 -3.299302 0.0048193554 0.0069696683697878879 0 +686 0 -3.299302 0.0048193554 0.0069696683697878879 0 +687 0 -2.4359035 0.014286714 0.020760022876868865 0 +690 0 -3.05244 0.0065817856 0.0095268959776332787 0 +695 0 -3.111528 0.006108972 0.0088404137863029985 0 diff --git a/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer-out.txt b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer-out.txt new file mode 100644 index 0000000000..078266882b --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer-out.txt @@ -0,0 +1,118 @@ +maml.exe TrainTest test=%Data% tr=WeightedEnsemble{bp=svm bp=ap nm=20 tp=-} dout=%Output% loader=Text{col=Label:BL:0 col=Features:R4:1-9} data=%Data% out=%Output% seed=1 +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Training 20 learners for the batch 1 +Beginning training model 1 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 26 instances with missing features during training (over 1 iterations; 26 inst/iter) +Trainer 1 of 20 finished in %Time% +Beginning training model 2 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 150 instances with missing features during training (over 10 iterations; 15 inst/iter) +Trainer 2 of 20 finished in %Time% +Beginning training model 3 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 11 instances with missing features during training (over 1 iterations; 11 inst/iter) +Trainer 3 of 20 finished in %Time% +Beginning training model 4 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 120 instances with missing features during training (over 10 iterations; 12 inst/iter) +Trainer 4 of 20 finished in %Time% +Beginning training model 5 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 18 instances with missing features during training (over 1 iterations; 18 inst/iter) +Trainer 5 of 20 finished in %Time% +Beginning training model 6 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 150 instances with missing features during training (over 10 iterations; 15 inst/iter) +Trainer 6 of 20 finished in %Time% +Beginning training model 7 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 24 instances with missing features during training (over 1 iterations; 24 inst/iter) +Trainer 7 of 20 finished in %Time% +Beginning training model 8 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 180 instances with missing features during training (over 10 iterations; 18 inst/iter) +Trainer 8 of 20 finished in %Time% +Beginning training model 9 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 15 instances with missing features during training (over 1 iterations; 15 inst/iter) +Trainer 9 of 20 finished in %Time% +Beginning training model 10 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 180 instances with missing features during training (over 10 iterations; 18 inst/iter) +Trainer 10 of 20 finished in %Time% +Beginning training model 11 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 14 instances with missing features during training (over 1 iterations; 14 inst/iter) +Trainer 11 of 20 finished in %Time% +Beginning training model 12 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 130 instances with missing features during training (over 10 iterations; 13 inst/iter) +Trainer 12 of 20 finished in %Time% +Beginning training model 13 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 14 instances with missing features during training (over 1 iterations; 14 inst/iter) +Trainer 13 of 20 finished in %Time% +Beginning training model 14 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 170 instances with missing features during training (over 10 iterations; 17 inst/iter) +Trainer 14 of 20 finished in %Time% +Beginning training model 15 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 24 instances with missing features during training (over 1 iterations; 24 inst/iter) +Trainer 15 of 20 finished in %Time% +Beginning training model 16 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 120 instances with missing features during training (over 10 iterations; 12 inst/iter) +Trainer 16 of 20 finished in %Time% +Beginning training model 17 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 12 instances with missing features during training (over 1 iterations; 12 inst/iter) +Trainer 17 of 20 finished in %Time% +Beginning training model 18 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 190 instances with missing features during training (over 10 iterations; 19 inst/iter) +Trainer 18 of 20 finished in %Time% +Beginning training model 19 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 23 instances with missing features during training (over 1 iterations; 23 inst/iter) +Trainer 19 of 20 finished in %Time% +Beginning training model 20 of 20 +Warning: Training data does not support shuffling, so ignoring request to shuffle +Warning: Skipped 180 instances with missing features during training (over 10 iterations; 18 inst/iter) +Trainer 20 of 20 finished in %Time% +Training calibrator. +Warning: The predictor produced non-finite prediction values on 16 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable. +TEST POSITIVE RATIO: 0.3499 (239.0/(239.0+444.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 231 | 8 | 0.9665 + negative || 9 | 435 | 0.9797 + ||====================== +Precision || 0.9625 | 0.9819 | +OVERALL 0/1 ACCURACY: 0.975110 +LOG LOSS/instance: 0.112863 +Test-set entropy (prior Log-Loss/instance): 0.934003 +LOG-LOSS REDUCTION (RIG): 0.879162 +AUC: 0.996249 + +OVERALL RESULTS +--------------------------------------- +AUC: 0.996249 (0.0000) +Accuracy: 0.975110 (0.0000) +Positive precision: 0.962500 (0.0000) +Positive recall: 0.966527 (0.0000) +Negative precision: 0.981941 (0.0000) +Negative recall: 0.979730 (0.0000) +Log-loss: 0.112863 (0.0000) +Log-loss reduction: 0.879162 (0.0000) +F1 Score: 0.964509 (0.0000) +AUPRC: 0.992435 (0.0000) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer-rp.txt b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer-rp.txt new file mode 100644 index 0000000000..a14165fac2 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer-rp.txt @@ -0,0 +1,4 @@ +WeightedEnsemble +AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC /bp /nm Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.996249 0.97511 0.9625 0.966527 0.981941 0.97973 0.112863 0.879162 0.964509 0.992435 svm,ap 20 WeightedEnsemble %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsemble{bp=svm bp=ap nm=20 tp=-} dout=%Output% loader=Text{col=Label:BL:0 col=Features:R4:1-9} data=%Data% out=%Output% seed=1 /bp:svm,ap;/nm:20 + diff --git a/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer.txt b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer.txt new file mode 100644 index 0000000000..b5583f73f1 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-Hetero-TrainTest-breast-cancer.txt @@ -0,0 +1,700 @@ +Instance Label Score Probability Log-loss Assigned +0 0 -3.2653844 0.01609207 0.023404773580146997 0 +1 0 2.3942997 0.8930092 3.2244411568032052 1 +2 0 -3.5103545 0.012332842 0.017903156367001596 0 +3 0 2.3300252 0.8860532 3.1335677427618682 1 +4 0 -2.9643283 0.02227979 0.032506420017895478 0 +5 1 9.490736 0.9999518 6.9568659593352019E-05 1 +6 0 -1.6140785 0.09161906 0.13863065882610109 0 +7 0 -3.8577213 0.008444489 0.012234552658491686 0 +8 0 -3.6269126 0.010862814 0.015757469294557003 0 +9 0 -3.6241956 0.010895023 0.015804447933838124 0 +10 0 -4.5891733 0.003790069 0.0054783017558651982 0 +11 0 -4.3108807 0.0051428624 0.0074387267544352214 0 +12 1 -0.49860764 0.25632605 1.9639479924952379 0 +13 0 -3.9268007 0.007830536 0.011341538789164538 0 +14 1 6.8671675 0.9991329 0.0012515445057059728 1 +15 1 1.0072587 0.64423734 0.63433581425753083 1 +16 0 -3.7142036 0.009876682 0.014319874162656562 0 +17 0 -3.5423803 0.011910398 0.017286219944511169 0 +18 1 6.187611 0.9981685 0.0026446780232283022 1 +19 0 -2.7876728 0.026937503 0.039395627305407972 0 +20 1 4.5410404 0.9888691 0.0161485780622295 1 +21 1 6.0276127 0.99781626 0.0031539090959622827 1 +22 0 -3.9911997 0.007298154 0.010567619464590899 0 +23 1 ? ? ? 0 +24 0 -4.561424 0.0039072633 0.0056480307271758777 0 +25 1 0.95733976 0.6315355 0.66306432895578327 1 +26 0 -3.9880342 0.0073234625 0.010604401090954271 0 +27 0 -3.4372077 0.013353979 0.019395513403520501 0 +28 0 -4.3108807 0.0051428624 0.0074387267544352214 0 +29 0 -4.175107 0.0059676734 0.0086353248944514086 0 +30 0 -4.12895 0.0062770187 0.0090843650252649174 0 +31 0 -4.180897 0.0059299525 0.0085805794795519568 0 +32 1 5.927128 0.9975613 0.0035226321919329248 1 +33 0 -4.0466037 0.006868991 0.0099440513860409134 0 +34 0 -3.6587865 0.010491926 0.015216616062819949 0 +35 0 -4.3108807 0.0051428624 0.0074387267544352214 0 +36 1 7.264369 0.99944 0.00080811478988956963 1 +37 0 -1.5557506 0.09710908 0.14737639041341738 0 +38 1 4.131503 0.982633 0.025275412463056989 1 +39 1 1.4290624 0.7424016 0.42972827539403025 1 +40 0 ? ? ? 0 +41 1 2.4566755 0.8993996 0.15296588687911028 1 +42 1 6.9449177 0.999204 0.0011488714741787922 1 +43 1 0.5260675 0.51591796 0.95478643367026872 1 +44 1 6.58985 0.9988234 0.0016984681652342905 1 +45 0 -4.5121536 0.0041243085 0.0059624232832799452 0 +46 1 3.703094 0.972447 0.040308504458262936 1 +47 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +48 0 -2.9643283 0.02227979 0.032506420017895478 0 +49 1 4.9304733 0.99272394 0.010535508213437971 1 +50 1 1.8948936 0.8280226 0.27225795115057777 1 +51 1 -0.28067923 0.30468872 1.7145919954765854 0 +52 1 4.19266 0.9837462 0.023641979610055389 1 +53 1 5.4079657 0.9956874 0.0062351849043310401 1 +54 1 4.8227596 0.9918147 0.011857526088524495 1 +55 1 4.216007 0.9841524 0.023046389058054482 1 +56 1 5.2737007 0.99500334 0.0072267226125532699 1 +57 1 1.0327678 0.6506519 0.62004225141698144 1 +58 1 1.3259258 0.7200817 0.47376751790611688 1 +59 1 1.067169 0.6592164 0.60117595185198103 1 +60 1 1.9315438 0.8336963 0.26240615243733373 1 +61 0 -4.3741302 0.0047983704 0.0069392470851821278 0 +62 1 5.711467 0.99690926 0.0044658992145244058 1 +63 1 0.13153958 0.40831316 1.2922520453284072 1 +64 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +65 1 2.4733293 0.90104747 0.1503249840629507 1 +66 0 -3.5423803 0.011910398 0.017286219944511169 0 +67 1 3.027186 0.9436982 0.083602501643435689 1 +68 1 7.6869254 0.99964833 0.00050743804998668485 1 +69 0 -4.360554 0.004870323 0.0070435569932352868 0 +70 0 -2.86837 0.024702694 0.036086023589653238 0 +71 1 5.8321104 0.9972929 0.0039108499875312938 1 +72 0 -2.5583076 0.034414854 0.050524612056300947 0 +73 1 6.377677 0.998514 0.0021454441981077982 1 +74 1 2.3740883 0.8908631 0.16672431349405054 1 +75 0 -3.6002493 0.0111830095 0.016224562077781109 0 +76 0 -3.7219477 0.009793601 0.014198822443033698 0 +77 0 -2.8432057 0.025379475 0.037087489937139485 0 +78 0 -3.2013326 0.017248312 0.025101158241811944 0 +79 0 -4.2326174 0.0056033605 0.0081066739524143661 0 +80 0 -3.066025 0.01996556 0.029095645011984261 0 +81 0 -3.5765634 0.011475266 0.016651030567526592 0 +82 0 -3.1724813 0.01779543 0.025904560498161484 0 +83 0 -2.616547 0.032345314 0.047435790225179117 0 +84 1 6.649848 0.99889857 0.0015899094538178553 1 +85 1 5.1913624 0.9945315 0.0079110105436784404 1 +86 1 1.3384473 0.7228537 0.46822436819631946 1 +87 1 5.110711 0.9940264 0.0086438939006461303 1 +88 0 -3.5423803 0.011910398 0.017286219944511169 0 +89 0 -4.098324 0.0064910194 0.0093950864237927739 0 +90 0 -4.561424 0.0039072633 0.0056480307271758777 0 +91 0 -4.0253825 0.0070303273 0.010178439338477008 0 +92 0 -3.5423803 0.011910398 0.017286219944511169 0 +93 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +94 0 -4.180897 0.0059299525 0.0085805794795519568 0 +95 0 -4.561424 0.0039072633 0.0056480307271758777 0 +96 0 -4.304512 0.0051788846 0.0074909653430758078 0 +97 0 -3.2653844 0.01609207 0.023404773580146997 0 +98 1 7.08919 0.99932086 0.00098011801603193282 1 +99 1 7.562501 0.9995967 0.00058193465852467608 1 +100 1 4.2504015 0.9847326 0.022196034085175179 1 +101 1 -0.70509744 0.21540724 2.2148613714634955 0 +102 0 -3.2099671 0.017087808 0.024865554872516737 0 +103 1 0.6878574 0.5601895 0.83601318981720607 1 +104 1 9.079937 0.9999242 0.00010938511263260336 1 +105 1 1.5024414 0.75755614 0.40057528813884447 1 +106 1 7.668572 0.9996412 0.00051776067608620972 1 +107 1 4.957925 0.9929391 0.010222838160957483 1 +108 0 -4.3402457 0.0049799667 0.0072025223073367553 0 +109 1 4.805126 0.99165547 0.012089123051715417 1 +110 0 -2.71747 0.029040622 0.042517155368045567 0 +111 1 3.0339081 0.9440904 0.083003045842630657 1 +112 1 5.526126 0.9962118 0.0054755564262458844 1 +113 1 7.3187904 0.99947256 0.00076113798427320249 1 +114 0 -2.6591587 0.030907841 0.045294225496711631 0 +115 0 -3.6688507 0.010377442 0.015049709292074737 0 +116 0 -0.6106701 0.2335081 0.38365753964720217 0 +117 1 6.245225 0.99828094 0.0024822099594683095 1 +118 0 -4.030241 0.0069930614 0.010124296315141848 0 +119 0 -3.4192936 0.013616517 0.019779453157641555 0 +120 0 -4.0978036 0.0064947163 0.0094004547572709724 0 +121 0 -3.1186867 0.018861435 0.027471194517187773 0 +122 1 7.9960036 0.99974984 0.00036095074256236041 1 +123 1 2.2949152 0.8820893 0.18100335016434929 1 +124 1 5.382405 0.9955648 0.0064128461295340847 1 +125 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +126 1 5.499481 0.9960994 0.0056383619525525027 1 +127 0 -3.8193765 0.008805644 0.012760121954867053 0 +128 1 4.0050287 0.98008823 0.029016458977127478 1 +129 0 -4.1953506 0.0058368254 0.0084454302880911086 0 +130 0 -2.86837 0.024702694 0.036086023589653238 0 +131 0 -4.180897 0.0059299525 0.0085805794795519568 0 +132 1 6.6597075 0.9989104 0.0015727784129403891 1 +133 0 -3.8269148 0.008733455 0.012655054314946195 0 +134 0 -3.8534493 0.008483987 0.01229202313515714 0 +135 0 -2.6051593 0.032740287 0.048024784653724777 0 +136 0 -3.7142036 0.009876682 0.014319874162656562 0 +137 0 -4.163023 0.0060471655 0.0087507008750277743 0 +138 0 -3.3817904 0.014182733 0.020607844796615989 0 +139 0 ? ? ? 0 +140 0 -4.163023 0.0060471655 0.0087507008750277743 0 +141 0 -4.4408646 0.004459773 0.0064484827624683733 0 +142 1 3.6463163 0.9707206 0.042872002786540617 1 +143 0 -3.6688507 0.010377442 0.015049709292074737 0 +144 0 -4.3108807 0.0051428624 0.0074387267544352214 0 +145 0 ? ? ? 0 +146 1 0.7448549 0.57559675 0.79686964924038517 1 +147 0 -4.2958665 0.005228187 0.0075624656598232351 0 +148 0 -1.4436777 0.10848597 0.16567058778935873 0 +149 1 9.159507 0.99993056 0.00010018337250504059 1 +150 0 -4.5891733 0.003790069 0.0054783017558651982 0 +151 1 3.6840632 0.9718796 0.041150494130796772 1 +152 1 7.6329145 0.9996268 0.00053849217329163485 1 +153 0 -3.2679832 0.016046803 0.023338401988860173 0 +154 0 -4.7929907 0.0030301402 0.004378204861445357 0 +155 1 2.7025304 0.92139256 0.11811214626904463 1 +156 0 -4.134472 0.0062391884 0.0090294437025751385 0 +157 0 -4.180897 0.0059299525 0.0085805794795519568 0 +158 0 ? ? ? 0 +159 1 10.04513 0.99997383 3.7750685767049689E-05 1 +160 1 7.5036116 0.9995697 0.00062090497120789344 1 +161 0 -3.3780956 0.014239756 0.02069129759088293 0 +162 0 -3.8193765 0.008805644 0.012760121954867053 0 +163 0 -3.1213365 0.01880749 0.027391873015351911 0 +164 0 ? ? ? 0 +165 0 -3.0690737 0.019899948 0.028999062212118336 0 +166 1 6.4891534 0.99868554 0.0018976141432561818 1 +167 1 6.233284 0.99825823 0.0025150294456999307 1 +168 0 -3.8193765 0.008805644 0.012760121954867053 0 +169 0 -4.6255016 0.0036419213 0.0052637727916244031 0 +170 0 -4.163023 0.0060471655 0.0087507008750277743 0 +171 0 -4.561424 0.0039072633 0.0056480307271758777 0 +172 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +173 1 10.937983 0.9999902 1.4102646298686894E-05 1 +174 1 4.192259 0.9837391 0.023652381688273109 1 +175 1 6.0337906 0.99783105 0.0031325367336339496 1 +176 0 -4.180897 0.0059299525 0.0085805794795519568 0 +177 1 4.079585 0.98162955 0.026749415104072465 1 +178 0 -3.5423803 0.011910398 0.017286219944511169 0 +179 1 1.358726 0.72730696 0.45936370858638231 1 +180 0 -4.5891733 0.003790069 0.0054783017558651982 0 +181 0 -4.7929907 0.0030301402 0.004378204861445357 0 +182 0 -2.7876728 0.026937503 0.039395627305407972 0 +183 1 6.992011 0.9992442 0.0010907822578806426 1 +184 1 4.7739973 0.99136686 0.012509057808797543 1 +185 0 -4.259161 0.005442758 0.0078736868831761463 0 +186 1 4.1567426 0.98310125 0.024588088764406185 1 +187 1 9.617186 0.99995804 6.0539163274785196E-05 1 +188 1 6.1444263 0.9980795 0.0027733905277275224 1 +189 0 -3.7622724 0.00937204 0.013584753523686708 0 +190 1 9.311988 0.9999413 8.4703942082841308E-05 1 +191 1 8.432388 0.99984527 0.00022325075346824142 1 +192 0 -3.4372077 0.013353979 0.019395513403520501 0 +193 0 -4.561424 0.0039072633 0.0056480307271758777 0 +194 0 -3.8193765 0.008805644 0.012760121954867053 0 +195 0 -3.5423803 0.011910398 0.017286219944511169 0 +196 0 4.806697 0.9916698 6.9074286576647337 1 +197 0 -2.4172082 0.039971277 0.058850524741158615 0 +198 0 -4.7929907 0.0030301402 0.004378204861445357 0 +199 0 -3.9911997 0.007298154 0.010567619464590899 0 +200 1 8.171349 0.99979377 0.00029756067032404669 1 +201 1 7.0345073 0.9992787 0.0010409564825537394 1 +202 0 -4.561424 0.0039072633 0.0056480307271758777 0 +203 0 -3.2653844 0.01609207 0.023404773580146997 0 +204 0 -4.561424 0.0039072633 0.0056480307271758777 0 +205 1 9.044386 0.99992114 0.00011377100942182498 1 +206 1 5.357063 0.9954398 0.0065939846417366766 1 +207 0 -4.5891733 0.003790069 0.0054783017558651982 0 +208 0 -4.5891733 0.003790069 0.0054783017558651982 0 +209 0 -3.3103275 0.015326635 0.022282860487917125 0 +210 1 10.73283 0.9999877 1.7714321792245208E-05 1 +211 1 7.4396763 0.9995383 0.0006662426199314735 1 +212 0 -4.561424 0.0039072633 0.0056480307271758777 0 +213 1 11.035078 0.9999912 1.272677229846128E-05 1 +214 1 10.208881 0.9999781 3.1559161611409118E-05 1 +215 1 6.245116 0.99828076 0.0024824683777035289 1 +216 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +217 0 -4.561424 0.0039072633 0.0056480307271758777 0 +218 1 6.0978603 0.99797857 0.0029192613497825144 1 +219 0 -2.263223 0.047013663 0.069472564205331072 0 +220 0 -4.3542194 0.0049042637 0.0070927634934080264 0 +221 1 7.3951235 0.9995151 0.00069970908489180958 1 +222 1 -2.0358088 0.059601173 4.06851547010963 0 +223 1 4.0871634 0.98177946 0.026529116446835269 1 +224 1 7.2424936 0.99942636 0.00082781797138124877 1 +225 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +226 1 7.48438 0.9995605 0.0006342394258747378 1 +227 1 6.233205 0.9982581 0.002515201728437437 1 +228 0 -4.5891733 0.003790069 0.0054783017558651982 0 +229 1 10.090591 0.9999751 3.5944821810158526E-05 1 +230 1 4.2729473 0.9851016 0.021655596409404121 1 +231 1 6.6903114 0.99894655 0.0015206117725055834 1 +232 0 0.964937 0.6334809 1.4480397385481536 1 +233 1 4.6214604 0.9898032 0.014786395779257951 1 +234 0 -2.2420673 0.048068956 0.071071022845288115 0 +235 0 ? ? ? 0 +236 1 9.714638 0.99996233 5.4347541318697156E-05 1 +237 1 5.7844896 0.9971475 0.0041211674736218009 1 +238 1 9.890707 0.99996895 4.4802176197198347E-05 1 +239 1 4.3297806 0.9859931 0.020350562853057669 1 +240 0 -1.683667 0.08543541 0.12884303720640797 0 +241 0 -3.3089218 0.015350024 0.022317129653216194 0 +242 0 -4.180897 0.0059299525 0.0085805794795519568 0 +243 0 -2.7772946 0.027238814 0.039842429956224272 0 +244 0 -4.561424 0.0039072633 0.0056480307271758777 0 +245 0 -2.5510187 0.034682695 0.050924852650500348 0 +246 1 8.931614 0.9999107 0.00012882075664009466 1 +247 1 2.6447172 0.916654 0.12555083220929728 1 +248 0 -2.703452 0.02947925 0.043169036866321266 0 +249 0 ? ? ? 0 +250 0 -4.2502556 0.0054961233 0.0079511000703902102 0 +251 1 5.7244315 0.99695295 0.0044026735400707721 1 +252 0 3.280855 0.9568324 4.5339076022407134 1 +253 1 6.9449177 0.999204 0.0011488714741787922 1 +254 1 5.711467 0.99690926 0.0044658992145244058 1 +255 1 3.4343557 0.963302 0.053939909438120381 1 +256 0 -4.163023 0.0060471655 0.0087507008750277743 0 +257 0 -3.9911997 0.007298154 0.010567619464590899 0 +258 0 -3.8193765 0.008805644 0.012760121954867053 0 +259 0 2.5125067 0.9048296 3.3933435781535293 1 +260 1 7.2856245 0.99945295 0.00078944433782944653 1 +261 1 9.371489 0.999945 7.9372176726815187E-05 1 +262 1 7.76371 0.9996769 0.00046623430328395426 1 +263 1 6.353957 0.9984747 0.0022021979324445707 1 +264 1 3.9517107 0.9789091 0.030753231752119768 1 +265 0 -2.3131795 0.044608466 0.065836002652573042 0 +266 1 6.232263 0.99825627 0.0025178721134996311 1 +267 1 1.6970851 0.79474354 0.33143871409780096 1 +268 1 7.2000847 0.99939895 0.00086739722772762864 1 +269 0 -4.561424 0.0039072633 0.0056480307271758777 0 +270 1 5.017771 0.9933867 0.0095726841193030775 1 +271 0 -3.2653844 0.01609207 0.023404773580146997 0 +272 1 1.6970851 0.79474354 0.33143871409780096 1 +273 1 -0.06442881 0.35736158 1.4845435367540205 0 +274 0 -3.6550915 0.010534272 0.015278356609565164 0 +275 0 ? ? ? 0 +276 0 -3.9911997 0.007298154 0.010567619464590899 0 +277 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +278 0 -4.561424 0.0039072633 0.0056480307271758777 0 +279 1 5.975487 0.9976875 0.003340068866493288 1 +280 0 -3.8193765 0.008805644 0.012760121954867053 0 +281 0 -4.0466037 0.006868991 0.0099440513860409134 0 +282 1 2.6876793 0.9201994 0.11998158818248605 1 +283 1 4.8468313 0.9920272 0.011548383441349655 1 +284 1 5.337451 0.99534065 0.0067377368711707885 1 +285 1 10.638188 0.99998635 1.9692147917725957E-05 1 +286 1 12.19091 0.99999756 3.5256486506625579E-06 1 +287 0 -3.8534493 0.008483987 0.01229202313515714 0 +288 1 0.85499 0.6049312 0.72515708161052861 1 +289 1 6.3008037 0.9983829 0.0023348329670084093 1 +290 0 -4.7929907 0.0030301402 0.004378204861445357 0 +291 0 -4.561424 0.0039072633 0.0056480307271758777 0 +292 1 ? ? ? 0 +293 1 4.320032 0.985844 0.020568698828840309 1 +294 0 ? ? ? 0 +295 1 5.3697605 0.9955029 0.0065025919345750296 1 +296 0 1.1073012 0.6690777 1.5954355561205926 1 +297 0 ? ? ? 0 +298 0 -1.8807185 0.06992958 0.10458813729619976 0 +299 1 4.882295 0.9923304 0.011107583975012837 1 +300 1 5.6863785 0.99682295 0.0045908060986382584 1 +301 0 -4.561424 0.0039072633 0.0056480307271758777 0 +302 1 12.254818 0.9999977 3.3536655897601792E-06 1 +303 0 -4.561424 0.0039072633 0.0056480307271758777 0 +304 1 4.039583 0.98081774 0.027943028739197025 1 +305 1 7.158686 0.99937093 0.00090783802421264394 1 +306 0 -4.561424 0.0039072633 0.0056480307271758777 0 +307 0 -4.561424 0.0039072633 0.0056480307271758777 0 +308 1 5.4057903 0.9956771 0.0062501259148902673 1 +309 0 -2.1225212 0.05446686 0.080800068514756962 0 +310 0 -4.2326174 0.0056033605 0.0081066739524143661 0 +311 0 -4.7929907 0.0030301402 0.004378204861445357 0 +312 1 2.7895415 0.9280609 0.10770863249926606 1 +313 0 -4.7929907 0.0030301402 0.004378204861445357 0 +314 0 -4.7049565 0.0033377076 0.0048233481554777833 0 +315 0 ? ? ? 0 +316 1 2.572311 0.9103538 0.13550078346044109 1 +317 1 6.9025755 0.999166 0.0012036926279508453 1 +318 0 -4.326701 0.005054454 0.0073105264714879107 0 +319 0 1.3834867 0.73268336 1.9033784523440218 1 +320 1 4.996194 0.9932287 0.0098021830035000222 1 +321 0 ? ? ? 0 +322 0 -3.8193765 0.008805644 0.012760121954867053 0 +323 1 3.6002264 0.9692424 0.045070586266708443 1 +324 0 -4.561424 0.0039072633 0.0056480307271758777 0 +325 0 -3.2402081 0.01653716 0.024057552907974622 0 +326 1 2.9760983 0.94063216 0.088297429563723301 1 +327 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +328 1 3.8758874 0.97711366 0.033401699359266249 1 +329 1 5.2012377 0.9945904 0.0078255864890170064 1 +330 1 4.254653 0.98480284 0.022093169442274737 1 +331 0 -2.941507 0.02283408 0.033324545396294918 0 +332 0 -2.5710278 0.033952214 0.049833540221445441 0 +333 1 3.4964824 0.96564627 0.050433293398985973 1 +334 1 5.0232086 0.9934259 0.009515726263910191 1 +335 0 -4.7929907 0.0030301402 0.004378204861445357 0 +336 1 4.2225 0.98426354 0.022883441968672362 1 +337 0 -4.561424 0.0039072633 0.0056480307271758777 0 +338 0 -4.7049565 0.0033377076 0.0048233481554777833 0 +339 1 4.4128485 0.98720235 0.018582271482583505 1 +340 1 4.5462866 0.98893255 0.016055969419739427 1 +341 0 -4.561424 0.0039072633 0.0056480307271758777 0 +342 0 -4.4408646 0.004459773 0.0064484827624683733 0 +343 0 -4.7929907 0.0030301402 0.004378204861445357 0 +344 1 6.766054 0.99903077 0.0013989830213576799 1 +345 0 -4.7929907 0.0030301402 0.004378204861445357 0 +346 0 -2.5197663 0.035854038 0.052676522158473391 0 +347 0 -4.540222 0.00399923 0.0057812372379503433 0 +348 1 -0.012272835 0.37066248 1.4318220056333359 0 +349 1 2.2752182 0.8798136 0.18473017367757963 1 +350 0 -3.0693142 0.019894779 0.028991453763807018 0 +351 0 -4.180897 0.0059299525 0.0085805794795519568 0 +352 0 -0.07698059 0.3541923 0.63082342285743431 0 +353 1 6.505441 0.99870884 0.0018639476740479471 1 +354 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +355 0 -3.4121685 0.013722348 0.01993425078263741 0 +356 1 -0.8419635 0.19101712 2.3882261404730252 0 +357 1 10.052368 0.999974 3.7492705063384392E-05 1 +358 1 5.0753384 0.99379057 0.0089862470228294668 1 +359 1 4.100359 0.9820377 0.026149650965844214 1 +360 1 11.307167 0.9999935 9.3730849200831842E-06 1 +361 1 4.063813 0.9813136 0.027213860570197302 1 +362 0 -2.7143378 0.029138075 0.042661963561225182 0 +363 0 -1.4885807 0.10379343 0.15809678783738457 0 +364 0 -4.180897 0.0059299525 0.0085805794795519568 0 +365 0 -4.3108807 0.0051428624 0.0074387267544352214 0 +366 1 10.271222 0.9999796 2.9409333050075951E-05 1 +367 1 7.986121 0.9997471 0.0003649073387411787 1 +368 0 -4.175107 0.0059676734 0.0086353248944514086 0 +369 0 -4.123387 0.006315364 0.0091400361796694484 0 +370 0 -3.0640237 0.020008745 0.029159219079037093 0 +371 0 -4.175107 0.0059676734 0.0086353248944514086 0 +372 0 -3.3817904 0.014182733 0.020607844796615989 0 +373 0 -3.2544837 0.016283318 0.023685227398143695 0 +374 0 -3.6587865 0.010491926 0.015216616062819949 0 +375 0 -4.7929907 0.0030301402 0.004378204861445357 0 +376 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +377 0 -4.7049565 0.0033377076 0.0048233481554777833 0 +378 0 -3.013068 0.021139657 0.030825054275845178 0 +379 0 -1.9886942 0.06257773 0.093229029017260295 0 +380 0 -4.7929907 0.0030301402 0.004378204861445357 0 +381 1 7.00312 0.9992534 0.0010775296384316602 1 +382 0 -3.0925539 0.01940165 0.028265762095443091 0 +383 0 -4.4408646 0.004459773 0.0064484827624683733 0 +384 0 -4.4408646 0.004459773 0.0064484827624683733 0 +385 0 -2.7200508 0.028960556 0.04239819490953757 0 +386 1 4.0645833 0.9813291 0.027191077263833848 1 +387 0 -2.0610461 0.058061793 0.086295676262700086 0 +388 0 -4.120144 0.006337824 0.0091726450893145888 0 +389 0 -2.8261507 0.025848394 0.037781780660790694 0 +390 0 -4.470003 0.0043194923 0.0062452078601454254 0 +391 1 8.021374 0.9997567 0.00035105929958808668 1 +392 0 -3.9911997 0.007298154 0.010567619464590899 0 +393 0 -4.8207397 0.0029391781 0.0042465815221570313 0 +394 0 -3.878056 0.008258964 0.011964642378794887 0 +395 0 -3.9911997 0.007298154 0.010567619464590899 0 +396 0 -3.8193765 0.008805644 0.012760121954867053 0 +397 0 -3.8860269 0.00818735 0.011860468506762941 0 +398 0 -3.7263813 0.009746348 0.01412997751942336 0 +399 0 -3.9938395 0.007277115 0.01053704371726912 0 +400 1 7.278433 0.9994486 0.0007957251541798663 1 +401 0 -4.163023 0.0060471655 0.0087507008750277743 0 +402 0 -2.4666376 0.037933223 0.055791059960573999 0 +403 0 -3.2359571 0.016613496 0.024169539830833704 0 +404 0 -3.9268205 0.0078303665 0.011341292321378862 0 +405 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +406 0 -3.2917824 0.015638039 0.022739185810068972 0 +407 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +408 0 -2.9637716 0.022293154 0.032526140382752297 0 +409 0 -3.6587865 0.010491926 0.015216616062819949 0 +410 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +411 0 ? ? ? 0 +412 1 7.5729628 0.99960136 0.00057522464469087896 1 +413 0 -2.9279583 0.023169508 0.033819860085709565 0 +414 1 5.3245487 0.9952743 0.0068338964475961579 1 +415 0 -0.90735507 0.18013191 0.28653628812146936 0 +416 1 6.484883 0.99867934 0.0019065690397204868 1 +417 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +418 0 -1.8329566 0.073430195 0.11002842690956885 0 +419 0 -3.8558304 0.008461949 0.012259957495037903 0 +420 0 -2.2032928 0.050061755 0.074094367084185597 0 +421 1 9.213476 0.99993455 9.4421565109537245E-05 1 +422 0 -2.23039 0.048661046 0.071968641793955263 0 +423 0 -2.86837 0.024702694 0.036086023589653238 0 +424 0 -4.163023 0.0060471655 0.0087507008750277743 0 +425 1 11.08436 0.99999166 1.2038835790403619E-05 1 +426 0 -1.8131723 0.07492698 0.11236084637713208 0 +427 1 3.7117805 0.9727022 0.039929906394138334 1 +428 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +429 0 -4.3108807 0.0051428624 0.0074387267544352214 0 +430 0 -3.9036996 0.008030756 0.011632704880047041 0 +431 0 -2.426865 0.039565027 0.058240156124209727 0 +432 0 -3.272923 0.015961103 0.023212751684584299 0 +433 0 -3.358839 0.014540629 0.021131703203387545 0 +434 0 3.9872332 0.97970206 5.6225225482656631 1 +435 1 6.3551836 0.9984768 0.0021991836415523884 1 +436 1 4.012264 0.9802432 0.028788357297963869 1 +437 0 -3.8860269 0.00818735 0.011860468506762941 0 +438 0 -2.9459198 0.022725863 0.033164783396578683 0 +439 0 -3.609031 0.011076534 0.016069221765043655 0 +440 1 5.512389 0.99615425 0.0055589423274722532 1 +441 0 -1.4009573 0.11312236 0.17319301942669602 0 +442 0 -3.8031878 0.008962676 0.012988702220263758 0 +443 0 -4.5041137 0.004160848 0.006015357938956213 0 +444 0 -2.2439585 0.047973715 0.070926688319408113 0 +445 0 -4.4408646 0.004459773 0.0064484827624683733 0 +446 0 -4.7929907 0.0030301402 0.004378204861445357 0 +447 0 -3.609031 0.011076534 0.016069221765043655 0 +448 0 -4.8207397 0.0029391781 0.0042465815221570313 0 +449 1 7.769636 0.99967897 0.00046322363723750687 1 +450 0 -3.2185173 0.01693032 0.024634415515610939 0 +451 0 -3.609031 0.011076534 0.016069221765043655 0 +452 0 -3.7536325 0.009460822 0.013714056519317612 0 +453 1 5.545363 0.9962909 0.0053610165703118166 1 +454 0 -3.9113288 0.007964077 0.011535731926001609 0 +455 1 0.8356905 0.59983873 0.73735340983669384 1 +456 1 7.411106 0.9995236 0.00068749244472971578 1 +457 1 6.881011 0.999146 0.0012326101197109947 1 +458 0 -3.2766178 0.015897298 0.023119210623573382 0 +459 0 -3.0443442 0.020438312 0.029791745104775773 0 +460 0 -3.0693142 0.019894779 0.028991453763807018 0 +461 0 -2.9281356 0.023165084 0.033813326552730699 0 +462 0 -2.5544052 0.034558006 0.050738513028946282 0 +463 0 -3.7217422 0.009795797 0.014202022024871176 0 +464 0 -3.8860269 0.00818735 0.011860468506762941 0 +465 1 7.5764675 0.99960285 0.00057307400583566552 1 +466 1 7.0032444 0.9992535 0.0010773575272915398 1 +467 1 5.7100124 0.9969043 0.0044730586401625586 1 +468 0 -3.8860269 0.00818735 0.011860468506762941 0 +469 0 -4.3682995 0.0048291427 0.006983856862971833 0 +470 0 -4.12895 0.0062770187 0.0090843650252649174 0 +471 0 -2.5544052 0.034558006 0.050738513028946282 0 +472 0 -3.2288501 0.016741898 0.024357925800841066 0 +473 0 -3.8860269 0.00818735 0.011860468506762941 0 +474 0 -3.609031 0.011076534 0.016069221765043655 0 +475 0 -4.163023 0.0060471655 0.0087507008750277743 0 +476 0 -3.5536137 0.011765635 0.017074869226522706 0 +477 0 -3.8860269 0.00818735 0.011860468506762941 0 +478 0 -3.5190227 0.012217063 0.017734047791872813 0 +479 1 5.7542105 0.997051 0.0042607924466669753 1 +480 0 -3.7467813 0.0095318165 0.013817462223198634 0 +481 0 -2.1517823 0.05283035 0.078305240690713637 0 +482 1 11.169056 0.99999243 1.0920939664449809E-05 1 +483 1 8.170057 0.99979347 0.00029799071570443488 1 +484 0 -3.2766178 0.015897298 0.023119210623573382 0 +485 0 -3.8735216 0.008299981 0.012024311878834735 0 +486 0 -4.12895 0.0062770187 0.0090843650252649174 0 +487 1 10.036007 0.99997354 3.8180653709006879E-05 1 +488 1 0.8369708 0.6001773 0.73653936946962328 1 +489 1 -0.6892204 0.21837807 2.1951001296339623 0 +490 0 -4.7929907 0.0030301402 0.004378204861445357 0 +491 1 5.010724 0.9933355 0.0096470443389943383 1 +492 0 -3.851954 0.008497857 0.012312203592380572 0 +493 1 7.7164946 0.9996596 0.00049118006365149263 1 +494 0 -0.08959174 0.3510208 0.62375588365287016 0 +495 0 -4.12895 0.0062770187 0.0090843650252649174 0 +496 0 -4.8207397 0.0029391781 0.0042465815221570313 0 +497 0 -3.7960186 0.009033103 0.01309122886919232 0 +498 0 -3.7142036 0.009876682 0.014319874162656562 0 +499 0 -3.7142036 0.009876682 0.014319874162656562 0 +500 0 -2.7876728 0.026937503 0.039395627305407972 0 +501 0 -3.7142036 0.009876682 0.014319874162656562 0 +502 0 -3.5765634 0.011475266 0.016651030567526592 0 +503 0 -3.5423803 0.011910398 0.017286219944511169 0 +504 0 -4.7929907 0.0030301402 0.004378204861445357 0 +505 0 -3.7836492 0.009155903 0.013270018500436942 0 +506 1 8.18573 0.999797 0.00029291618838410085 1 +507 0 -3.9289443 0.0078122094 0.011314890719336414 0 +508 0 -3.609031 0.011076534 0.016069221765043655 0 +509 0 -4.4408646 0.004459773 0.0064484827624683733 0 +510 0 -4.7929907 0.0030301402 0.004378204861445357 0 +511 0 -3.4372077 0.013353979 0.019395513403520501 0 +512 0 -3.609031 0.011076534 0.016069221765043655 0 +513 0 -4.12895 0.0062770187 0.0090843650252649174 0 +514 1 7.3587437 0.99949527 0.00072835837215675541 1 +515 1 6.1920147 0.99817735 0.0026319280120138842 1 +516 0 -4.8207397 0.0029391781 0.0042465815221570313 0 +517 0 -4.7049565 0.0033377076 0.0048233481554777833 0 +518 0 -3.6929693 0.010108089 0.014657093013472967 0 +519 1 4.455138 0.98777777 0.017741594381557761 1 +520 0 -4.7431955 0.0032004628 0.0046246963287338511 0 +521 0 -3.7482767 0.0095162755 0.013794825737778479 0 +522 1 3.58242 0.9686522 0.04594936072464742 1 +523 1 5.427267 0.9957778 0.0061042633610080711 1 +524 0 -3.9911997 0.007298154 0.010567619464590899 0 +525 0 -4.0253825 0.0070303273 0.010178439338477008 0 +526 0 -3.8860269 0.00818735 0.011860468506762941 0 +527 0 -3.5423803 0.011910398 0.017286219944511169 0 +528 0 -2.3386855 0.043426134 0.064051717595393159 0 +529 0 -3.851954 0.008497857 0.012312203592380572 0 +530 1 4.8951244 0.9924372 0.010952304879643498 1 +531 0 -3.2917824 0.015638039 0.022739185810068972 0 +532 0 -4.5891733 0.003790069 0.0054783017558651982 0 +533 0 -3.9911997 0.007298154 0.010567619464590899 0 +534 0 -4.3108807 0.0051428624 0.0074387267544352214 0 +535 0 -3.780846 0.009183961 0.013310872198908509 0 +536 0 -3.2653844 0.01609207 0.023404773580146997 0 +537 0 -2.9279583 0.023169508 0.033819860085709565 0 +538 0 -3.7142036 0.009876682 0.014319874162656562 0 +539 0 -3.0149083 0.021097744 0.030763281694029396 0 +540 0 -3.1281857 0.018668748 0.02718788982033506 0 +541 0 -4.163023 0.0060471655 0.0087507008750277743 0 +542 0 -3.4282966 0.013483945 0.019585565098181677 0 +543 0 -3.7142036 0.009876682 0.014319874162656562 0 +544 0 -3.5302353 0.01206889 0.01751765103006081 0 +545 0 -3.4372077 0.013353979 0.019395513403520501 0 +546 1 8.759241 0.99989206 0.00015573869603731518 1 +547 0 -4.5777693 0.0038378015 0.0055474290028238501 0 +548 0 -4.300773 0.005200149 0.0075218034551329058 0 +549 1 4.5371666 0.98882204 0.016217190552116956 1 +550 0 -3.9911997 0.007298154 0.010567619464590899 0 +551 0 -4.561424 0.0039072633 0.0056480307271758777 0 +552 0 -2.9232063 0.023288285 0.033995294446328962 0 +553 0 -1.3548462 0.11831982 0.18167266122486811 0 +554 0 -4.163023 0.0060471655 0.0087507008750277743 0 +555 0 -1.7657175 0.078632094 0.11815074940020485 0 +556 0 -2.6747835 0.030396404 0.044533045117446393 0 +557 0 -3.0693142 0.019894779 0.028991453763807018 0 +558 0 -4.3108807 0.0051428624 0.0074387267544352214 0 +559 0 -3.4372077 0.013353979 0.019395513403520501 0 +560 0 -3.2653844 0.01609207 0.023404773580146997 0 +561 0 -3.2653844 0.01609207 0.023404773580146997 0 +562 0 -4.561424 0.0039072633 0.0056480307271758777 0 +563 0 -3.9911997 0.007298154 0.010567619464590899 0 +564 0 -3.3780956 0.014239756 0.02069129759088293 0 +565 1 8.53931 0.9998625 0.0001983956285675477 1 +566 0 -3.4869633 0.012650708 0.018367541036486416 0 +567 0 -3.09616 0.019326214 0.028154780279773872 0 +568 1 3.6918535 0.97211325 0.040803696562312378 1 +569 1 7.642125 0.9996306 0.0005330727075031368 1 +570 1 5.45413 0.9959004 0.0059266401299438087 1 +571 1 8.001436 0.9997513 0.00035880042310327935 1 +572 0 -3.9911997 0.007298154 0.010567619464590899 0 +573 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +574 1 5.27071 0.99498695 0.0072504891752000443 1 +575 0 -2.9279583 0.023169508 0.033819860085709565 0 +576 0 -3.4372077 0.013353979 0.019395513403520501 0 +577 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +578 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +579 0 -4.561424 0.0039072633 0.0056480307271758777 0 +580 0 -3.1047945 0.01914675 0.0278907902198956 0 +581 1 6.6856213 0.99894106 0.0015285313390463744 1 +582 1 6.2932367 0.9983694 0.002354384746826962 1 +583 0 -4.163023 0.0060471655 0.0087507008750277743 0 +584 0 -2.3569767 0.042596724 0.062801352097153817 0 +585 0 -4.7929907 0.0030301402 0.004378204861445357 0 +586 1 10.1500225 0.9999767 3.3623000044204345E-05 1 +587 0 -3.272923 0.015961103 0.023212751684584299 0 +588 1 4.3938894 0.98693573 0.018971949595457907 1 +589 0 -3.609031 0.011076534 0.016069221765043655 0 +590 1 2.3242974 0.88541454 0.1755750285205338 1 +591 1 4.000958 0.98000056 0.029145527864694164 1 +592 1 4.0920124 0.98187476 0.026389071292179184 1 +593 0 -3.2766178 0.015897298 0.023119210623573382 0 +594 1 3.4444263 0.9636922 0.053355684303819866 1 +595 0 -3.4372077 0.013353979 0.019395513403520501 0 +596 0 -3.3817904 0.014182733 0.020607844796615989 0 +597 0 -2.6891704 0.029932741 0.04384331588399755 0 +598 0 -3.9911997 0.007298154 0.010567619464590899 0 +599 0 -2.5371537 0.03519773 0.051694794908747274 0 +600 0 -3.9911997 0.007298154 0.010567619464590899 0 +601 0 -4.7049565 0.0033377076 0.0048233481554777833 0 +602 0 -3.7142036 0.009876682 0.014319874162656562 0 +603 1 3.307713 0.95803815 0.061844986375055984 1 +604 1 2.964779 0.9399319 0.089371815851646261 1 +605 1 7.7185054 0.9996604 0.0004900617961935143 1 +606 0 -3.7483866 0.009515135 0.013793163997434152 0 +607 0 -4.7929907 0.0030301402 0.004378204861445357 0 +608 1 7.9886675 0.99974775 0.00036396119518952627 1 +609 0 -3.609031 0.011076534 0.016069221765043655 0 +610 1 5.3112736 0.995205 0.0069343827093675797 1 +611 1 5.178556 0.99445426 0.0080230724388507223 1 +612 1 11.885635 0.99999654 4.987505496102727E-06 1 +613 0 -3.8232942 0.008768054 0.012705410244507881 0 +614 0 -4.405946 0.0046338667 0.0067007938425663483 0 +615 0 -3.0885918 0.019484866 0.028388197273568373 0 +616 0 -3.9911997 0.007298154 0.010567619464590899 0 +617 0 ? ? ? 0 +618 0 -3.7142036 0.009876682 0.014319874162656562 0 +619 0 -3.4372077 0.013353979 0.019395513403520501 0 +620 0 -3.9911997 0.007298154 0.010567619464590899 0 +621 0 -0.21737361 0.31966156 0.55567548441110171 0 +622 0 -1.8795394 0.070014104 0.10471925853942105 0 +623 0 -4.7929907 0.0030301402 0.004378204861445357 0 +624 0 -3.1539073 0.018156638 0.026435212326140302 0 +625 0 -2.4943914 0.036833026 0.054142169737077465 0 +626 1 4.1199784 0.982415 0.025595476172414866 1 +627 0 -3.2984931 0.015524642 0.022572999247524317 0 +628 0 -4.4408646 0.004459773 0.0064484827624683733 0 +629 0 -3.8860269 0.00818735 0.011860468506762941 0 +630 0 -2.458294 0.038270105 0.056296329159577795 0 +631 0 -3.4372077 0.013353979 0.019395513403520501 0 +632 0 -4.7929907 0.0030301402 0.004378204861445357 0 +633 1 3.809505 0.97541994 0.035904632074950059 1 +634 0 -4.163023 0.0060471655 0.0087507008750277743 0 +635 0 -3.2250643 0.016810693 0.024458869372985807 0 +636 1 7.4516163 0.9995443 0.00065755351047660116 1 +637 0 -1.9161325 0.067434296 0.10072272074904487 0 +638 0 -3.8860269 0.00818735 0.011860468506762941 0 +639 0 -3.0693142 0.019894779 0.028991453763807018 0 +640 0 -3.4302614 0.013455182 0.019543502342225409 0 +641 0 -3.9911997 0.007298154 0.010567619464590899 0 +642 0 -3.9911997 0.007298154 0.010567619464590899 0 +643 0 -4.7929907 0.0030301402 0.004378204861445357 0 +644 0 -4.4408646 0.004459773 0.0064484827624683733 0 +645 0 -3.9911997 0.007298154 0.010567619464590899 0 +646 0 -4.2502556 0.0054961233 0.0079511000703902102 0 +647 0 -4.3785825 0.004775005 0.0069053756819761388 0 +648 1 8.495775 0.9998557 0.00020820002115843166 1 +649 0 -3.9911997 0.007298154 0.010567619464590899 0 +650 0 -2.9605665 0.022370243 0.032639897107013247 0 +651 0 -3.97648 0.0074165817 0.01073974118593138 0 +652 0 -3.272923 0.015961103 0.023212751684584299 0 +653 0 -3.7142036 0.009876682 0.014319874162656562 0 +654 0 -3.8193765 0.008805644 0.012760121954867053 0 +655 0 -3.9911997 0.007298154 0.010567619464590899 0 +656 0 -3.4372077 0.013353979 0.019395513403520501 0 +657 0 -0.42248583 0.27263552 0.45924961872848086 0 +658 1 6.8514137 0.9991177 0.0012734914913334666 1 +659 0 -4.7929907 0.0030301402 0.004378204861445357 0 +660 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +661 0 -3.5423803 0.011910398 0.017286219944511169 0 +662 0 -4.045123 0.006880127 0.0099602281925481612 0 +663 0 -4.045123 0.006880127 0.0099602281925481612 0 +664 0 -3.520666 0.012195237 0.017702169735039734 0 +665 0 -4.7929907 0.0030301402 0.004378204861445357 0 +666 0 -2.6617477 0.030822525 0.045167219603029475 0 +667 0 -3.8193765 0.008805644 0.012760121954867053 0 +668 1 2.0417037 0.8498484 0.23472260310098952 1 +669 1 6.798136 0.9990644 0.001350437678126858 1 +670 1 4.534569 0.98879033 0.016263455821978415 1 +671 0 -3.471818 0.012860818 0.018674582380787411 0 +672 0 -4.180897 0.0059299525 0.0085805794795519568 0 +673 0 -2.9305816 0.023104189 0.033723392383293277 0 +674 0 -4.6772075 0.0034409605 0.0049728172606640694 0 +675 0 -3.0559897 0.020183032 0.029415819707341133 0 +676 0 -4.3682995 0.0048291427 0.006983856862971833 0 +677 0 -3.609031 0.011076534 0.016069221765043655 0 +678 0 -4.7929907 0.0030301402 0.004378204861445357 0 +679 0 -4.4408646 0.004459773 0.0064484827624683733 0 +680 1 12.114309 0.9999973 3.8696148339734104E-06 1 +681 1 7.529873 0.99958193 0.00060326926884453412 1 +682 0 -3.1010995 0.019223347 0.028003458684784346 0 +683 0 -4.7929907 0.0030301402 0.004378204861445357 0 +684 0 -4.7929907 0.0030301402 0.004378204861445357 0 +685 0 -4.7929907 0.0030301402 0.004378204861445357 0 +686 0 -4.7929907 0.0030301402 0.004378204861445357 0 +687 0 -3.6626298 0.01044806 0.015152660867682972 0 +688 0 -3.8860269 0.00818735 0.011860468506762941 0 +689 0 -3.6151924 0.011002428 0.015961115582657694 0 +690 0 -4.3785825 0.004775005 0.0069053756819761388 0 +691 1 3.916963 0.97810405 0.031940141424689809 1 +692 0 -4.163023 0.0060471655 0.0087507008750277743 0 +693 0 -3.8830423 0.008214092 0.011899368056717551 0 +694 0 -3.7162507 0.009854654 0.014287776989780327 0 +695 0 -4.4408646 0.004459773 0.0064484827624683733 0 +696 1 5.3014183 0.9951529 0.0070099032168673631 1 +697 1 3.0674176 0.9460074 0.0800766703186329 1 +698 1 3.9357042 0.9785419 0.031294452528203641 1 diff --git a/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-RandomFeature-TrainTest-breast-cancer-out.txt b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-RandomFeature-TrainTest-breast-cancer-out.txt new file mode 100644 index 0000000000..09b1d929e7 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-RandomFeature-TrainTest-breast-cancer-out.txt @@ -0,0 +1,94 @@ +maml.exe TrainTest test=%Data% tr=WeightedEnsemble{nm=20 st=AllInstanceSelector{fs=RandomFeatureSelector} tp=-} dout=%Output% loader=Text{col=Label:BL:0 col=Features:R4:1-9} data=%Data% out=%Output% seed=1 +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Training 20 learners for the batch 1 +Beginning training model 1 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 1 of 20 finished in %Time% +Beginning training model 2 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 2 of 20 finished in %Time% +Beginning training model 3 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 3 of 20 finished in %Time% +Beginning training model 4 of 20 +Trainer 4 of 20 finished in %Time% +Beginning training model 5 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 5 of 20 finished in %Time% +Beginning training model 6 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 6 of 20 finished in %Time% +Beginning training model 7 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 7 of 20 finished in %Time% +Beginning training model 8 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 8 of 20 finished in %Time% +Beginning training model 9 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 9 of 20 finished in %Time% +Beginning training model 10 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 10 of 20 finished in %Time% +Beginning training model 11 of 20 +Trainer 11 of 20 finished in %Time% +Beginning training model 12 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 12 of 20 finished in %Time% +Beginning training model 13 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 13 of 20 finished in %Time% +Beginning training model 14 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 14 of 20 finished in %Time% +Beginning training model 15 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 15 of 20 finished in %Time% +Beginning training model 16 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 16 of 20 finished in %Time% +Beginning training model 17 of 20 +Trainer 17 of 20 finished in %Time% +Beginning training model 18 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 18 of 20 finished in %Time% +Beginning training model 19 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 19 of 20 finished in %Time% +Beginning training model 20 of 20 +Warning: Skipped 16 instances with missing features during training (over 1 iterations; 16 inst/iter) +Trainer 20 of 20 finished in %Time% +Training calibrator. +TEST POSITIVE RATIO: 0.3448 (241.0/(241.0+458.0)) +Confusion table + ||====================== +PREDICTED || positive | negative | Recall +TRUTH ||====================== + positive || 233 | 8 | 0.9668 + negative || 12 | 446 | 0.9738 + ||====================== +Precision || 0.9510 | 0.9824 | +OVERALL 0/1 ACCURACY: 0.971388 +LOG LOSS/instance: 0.130701 +Test-set entropy (prior Log-Loss/instance): 0.929318 +LOG-LOSS REDUCTION (RIG): 0.859358 +AUC: 0.994931 + +OVERALL RESULTS +--------------------------------------- +AUC: 0.994931 (0.0000) +Accuracy: 0.971388 (0.0000) +Positive precision: 0.951020 (0.0000) +Positive recall: 0.966805 (0.0000) +Negative precision: 0.982379 (0.0000) +Negative recall: 0.973799 (0.0000) +Log-loss: 0.130701 (0.0000) +Log-loss reduction: 0.859358 (0.0000) +F1 Score: 0.958848 (0.0000) +AUPRC: 0.989467 (0.0000) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-RandomFeature-TrainTest-breast-cancer-rp.txt b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-RandomFeature-TrainTest-breast-cancer-rp.txt new file mode 100644 index 0000000000..bc16118eb7 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsemble/net9.0/WE-RandomFeature-TrainTest-breast-cancer-rp.txt @@ -0,0 +1,4 @@ +WeightedEnsemble +AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC /bp /nm /st Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.994931 0.971388 0.95102 0.966805 0.982379 0.973799 0.130701 0.859358 0.958848 0.989467 Microsoft.ML.Runtime.ComponentFactoryUtils+SimpleComponentFactory`1[Microsoft.ML.Trainers.LinearSvmTrainer] 20 AllInstanceSelector{fs=RandomFeatureSelector} WeightedEnsemble %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsemble{nm=20 st=AllInstanceSelector{fs=RandomFeatureSelector} tp=-} dout=%Output% loader=Text{col=Label:BL:0 col=Features:R4:1-9} data=%Data% out=%Output% seed=1 /bp:Microsoft.ML.Runtime.ComponentFactoryUtils+SimpleComponentFactory`1[Microsoft.ML.Trainers.LinearSvmTrainer];/nm:20;/st:AllInstanceSelector{fs=RandomFeatureSelector} + diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris-out.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris-out.txt new file mode 100644 index 0000000000..77d4c1ab5c --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris-out.txt @@ -0,0 +1,61 @@ +maml.exe TrainTest test=%Data% tr=WeightedEnsembleMulticlass{bp=mlr{t-} nm=5 oc=MultiAverage tp=-} dout=%Output% data=%Data% out=%Output% seed=1 xf=Term{col=Label} +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Training 5 learners for the batch 1 +Beginning training model 1 of 5 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 12 of 15 weights. +Trainer 1 of 5 finished in %Time% +Beginning training model 2 of 5 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 10 of 15 weights. +Trainer 2 of 5 finished in %Time% +Beginning training model 3 of 5 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 3 of 5 finished in %Time% +Beginning training model 4 of 5 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 10 of 15 weights. +Trainer 4 of 5 finished in %Time% +Beginning training model 5 of 5 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 5 of 5 finished in %Time% +Not training a calibrator because it is not needed. + +Confusion table + ||======================== +PREDICTED || 0 | 1 | 2 | Recall +TRUTH ||======================== + 0 || 50 | 0 | 0 | 1.0000 + 1 || 0 | 45 | 5 | 0.9000 + 2 || 0 | 3 | 47 | 0.9400 + ||======================== +Precision ||1.0000 |0.9375 |0.9038 | +Accuracy(micro-avg): 0.946667 +Accuracy(macro-avg): 0.946667 +Log-loss: 0.433386 +Log-loss reduction: 0.605515 + +OVERALL RESULTS +--------------------------------------- +Accuracy(micro-avg): 0.946667 (0.0000) +Accuracy(macro-avg): 0.946667 (0.0000) +Log-loss: 0.433386 (0.0000) +Log-loss reduction: 0.605515 (0.0000) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris-rp.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris-rp.txt new file mode 100644 index 0000000000..e80183ddd5 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris-rp.txt @@ -0,0 +1,4 @@ +WeightedEnsembleMulticlass +Accuracy(micro-avg) Accuracy(macro-avg) Log-loss Log-loss reduction /oc /bp /nm Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.946667 0.946667 0.433386 0.605515 MultiAverage mlr{t-} 5 WeightedEnsembleMulticlass %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsembleMulticlass{bp=mlr{t-} nm=5 oc=MultiAverage tp=-} dout=%Output% data=%Data% out=%Output% seed=1 xf=Term{col=Label} /oc:MultiAverage;/bp:mlr{t-};/nm:5 + diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris.txt new file mode 100644 index 0000000000..f9acac3e28 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Average-TrainTest-iris.txt @@ -0,0 +1,151 @@ +Instance Label Assigned Log-loss #1 Score #2 Score #3 Score #1 Class #2 Class #3 Class +0 0 0 0.20109623732391574 0.8178337 0.1707502 0.011416035 0 1 2 +1 0 0 0.23100949707007967 0.7937319 0.19445382 0.01181429 0 1 2 +2 0 0 0.20993058590249727 0.8106405 0.17841254 0.010947041 0 1 2 +3 0 0 0.23309455644514918 0.7920787 0.19549197 0.012429379 0 1 2 +4 0 0 0.19544659347094889 0.82246727 0.16621368 0.011319163 0 1 2 +5 0 0 0.25812900770421759 0.77249557 0.20625499 0.021249486 0 1 2 +6 0 0 0.23239857723143698 0.79263014 0.19288896 0.014480923 0 1 2 +7 0 0 0.21488006167314444 0.8066382 0.1811498 0.012212071 0 1 2 +8 0 0 0.23693414633080512 0.78904325 0.19911228 0.011844534 0 1 2 +9 0 0 0.20743493257440657 0.8126661 0.17747913 0.009854694 0 1 2 +10 0 0 0.19810014115346058 0.8202877 0.16772966 0.011982614 0 1 2 +11 0 0 0.22313660739720248 0.80000556 0.18704335 0.012951198 0 1 2 +12 0 0 0.20518219978417895 0.8144989 0.17616834 0.009332846 0 1 2 +13 0 0 0.18211410646572782 0.8335062 0.158764 0.007729848 0 1 2 +14 0 0 0.16216736597316558 0.8502989 0.13995433 0.00974694 0 1 2 +15 0 0 0.20920365985465014 0.81123 0.17051752 0.01825247 0 1 2 +16 0 0 0.22087101129821127 0.8018201 0.18144083 0.016739113 0 1 2 +17 0 0 0.22669855951157375 0.79716104 0.18839757 0.01444138 0 1 2 +18 0 0 0.23544944486258634 0.7902156 0.19276467 0.01701977 0 1 2 +19 0 0 0.21708486133430296 0.80486166 0.18012397 0.01501447 0 1 2 +20 0 0 0.2326729392118223 0.7924127 0.19376366 0.013823678 0 1 2 +21 0 0 0.25159669610773255 0.77755827 0.20336466 0.019077076 0 1 2 +22 0 0 0.1666053894597781 0.8465336 0.14465752 0.008808913 0 1 2 +23 0 0 0.33960050330284824 0.7120547 0.26049328 0.027452001 0 1 2 +24 0 0 0.25048725365591912 0.7784214 0.20608087 0.015497719 0 1 2 +25 0 0 0.2495255993252532 0.77917033 0.20750463 0.013324998 0 1 2 +26 0 0 0.28315166840818462 0.7534055 0.22600721 0.020587329 0 1 2 +27 0 0 0.20920733358182858 0.811227 0.17662928 0.012143728 0 1 2 +28 0 0 0.20690816397004042 0.8130943 0.17539372 0.011511907 0 1 2 +29 0 0 0.23572306084889846 0.7899994 0.19688119 0.013119489 0 1 2 +30 0 0 0.24246643906046669 0.7846901 0.20209233 0.013217597 0 1 2 +31 0 0 0.27311834766781729 0.7610027 0.21951385 0.019483557 0 1 2 +32 0 0 0.15665831909463507 0.85499614 0.13589051 0.009113441 0 1 2 +33 0 0 0.16577454167174635 0.8472372 0.14193675 0.010826165 0 1 2 +34 0 0 0.20743493257440657 0.8126661 0.17747913 0.009854694 0 1 2 +35 0 0 0.20225447879485656 0.816887 0.17278682 0.01032619 0 1 2 +36 0 0 0.19380481743123668 0.8238187 0.16540623 0.010775076 0 1 2 +37 0 0 0.20743493257440657 0.8126661 0.17747913 0.009854694 0 1 2 +38 0 0 0.22170831843100872 0.801149 0.18777351 0.011077426 0 1 2 +39 0 0 0.21499511917116837 0.8065454 0.18123226 0.012222429 0 1 2 +40 0 0 0.21794146084390165 0.8041725 0.1822428 0.0135847 0 1 2 +41 0 0 0.30246474806932394 0.73899454 0.24643934 0.01456619 0 1 2 +42 0 0 0.20959675171676889 0.8109112 0.17816941 0.01091937 0 1 2 +43 0 0 0.34910900149310742 0.70531625 0.2626272 0.032056607 0 1 2 +44 0 0 0.2858108873265568 0.7514047 0.22462244 0.023972863 0 1 2 +45 0 0 0.25966440661143864 0.7713104 0.21380298 0.014886627 0 1 2 +46 0 0 0.20005312379213039 0.81868726 0.16870461 0.012608141 0 1 2 +47 0 0 0.2181078723528164 0.8040387 0.18433964 0.01162168 0 1 2 +48 0 0 0.19799180634473759 0.8203766 0.16765115 0.011972292 0 1 2 +49 0 0 0.2125455637460732 0.8085235 0.17989932 0.01157721 0 1 2 +50 1 1 0.7218109833856774 0.48587155 0.3959195 0.118208945 1 2 0 +51 1 1 0.74982489681486753 0.47244927 0.41691938 0.110631384 1 2 0 +52 1 1 0.76958688849763301 0.46320438 0.44827634 0.0885192 1 2 0 +53 1 1 0.57011251107542671 0.5654618 0.27743372 0.1571045 1 2 0 +54 1 1 0.71145945222736506 0.4909272 0.4124599 0.09661289 1 2 0 +55 1 1 0.63369918723000795 0.5306253 0.32881245 0.14056228 1 2 0 +56 1 2 0.82670132538770302 0.47655436 0.43749005 0.08595567 2 1 0 +57 1 1 0.60693117621784676 0.5450209 0.30628267 0.14869651 1 0 2 +58 1 1 0.64857824730524838 0.5227885 0.34067783 0.13653363 1 2 0 +59 1 1 0.63221148491088719 0.5314153 0.31361192 0.15497288 1 2 0 +60 1 1 0.54290968199341938 0.5810551 0.26368746 0.15525745 1 0 2 +61 1 1 0.7079825463638928 0.49263707 0.38454008 0.12282284 1 2 0 +62 1 1 0.53680433039379738 0.5846135 0.22836316 0.18702339 1 0 2 +63 1 1 0.68709468166099852 0.5030354 0.38551295 0.111451745 1 2 0 +64 1 1 0.63788071795847689 0.5284111 0.26065055 0.21093841 1 2 0 +65 1 1 0.69452696627418598 0.49931058 0.36718538 0.13350406 1 2 0 +66 1 1 0.72590845515241487 0.48388478 0.40953875 0.1065765 1 2 0 +67 1 1 0.58924895331247251 0.55474377 0.2447762 0.20048003 1 0 2 +68 1 1 0.64019658220479425 0.5271888 0.38371113 0.08910015 1 2 0 +69 1 1 0.57338754335302056 0.56361294 0.22376128 0.21262579 1 0 2 +70 1 2 0.94933591372849113 0.5579522 0.38699794 0.055049848 2 1 0 +71 1 1 0.62436164346709877 0.5356032 0.28985795 0.17453885 1 2 0 +72 1 1 0.70197701894644582 0.49560452 0.42634603 0.07804952 1 2 0 +73 1 1 0.61375751281993929 0.54131305 0.30953977 0.14914724 1 2 0 +74 1 1 0.63983722493158113 0.52737826 0.31627303 0.15634872 1 2 0 +75 1 1 0.68308898646095273 0.5050545 0.3641959 0.13074952 1 2 0 +76 1 1 0.68296200901721571 0.5051186 0.39143226 0.10344916 1 2 0 +77 1 2 0.88368574090766316 0.52964145 0.41325694 0.057101667 2 1 0 +78 1 1 0.71504523221102501 0.48916999 0.40680262 0.10402749 1 2 0 +79 1 1 0.61503506823226572 0.54062194 0.29691574 0.16246234 1 0 2 +80 1 1 0.56582147382446679 0.56789345 0.22770098 0.2044057 1 0 2 +81 1 1 0.57615855317640152 0.5620533 0.2663304 0.17161632 1 0 2 +82 1 1 0.60064200261137768 0.5484594 0.2467321 0.20480862 1 2 0 +83 1 2 0.79645591921191894 0.48746848 0.45092425 0.06160729 2 1 0 +84 1 1 0.72569032819821455 0.48399034 0.40928945 0.10672028 1 2 0 +85 1 2 0.82031002688932086 0.4623315 0.44029513 0.09737339 2 1 0 +86 1 1 0.75307737376463535 0.47091514 0.43111873 0.09796616 1 2 0 +87 1 1 0.57845925847006174 0.5607617 0.30808374 0.13115464 1 2 0 +88 1 1 0.64780189275442868 0.52319455 0.3017528 0.17505264 1 2 0 +89 1 1 0.59123027328785327 0.55364573 0.28220195 0.16415234 1 2 0 +90 1 1 0.58739970298115685 0.5557706 0.28101826 0.16321114 1 2 0 +91 1 1 0.69258073926487607 0.5002833 0.38027725 0.11943948 1 2 0 +92 1 1 0.58792985743135939 0.555476 0.25198236 0.19254175 1 2 0 +93 1 1 0.5941354540121756 0.5520396 0.3001401 0.14782034 1 0 2 +94 1 1 0.61548896842370293 0.5403766 0.3024408 0.15718262 1 2 0 +95 1 1 0.63160195763108873 0.5317393 0.27497527 0.19328542 1 2 0 +96 1 1 0.63770420191217425 0.5285044 0.30747172 0.16402392 1 2 0 +97 1 1 0.63974658646147831 0.52742606 0.31603643 0.15653752 1 2 0 +98 1 1 0.62807521349934403 0.5336179 0.3095249 0.15685731 1 0 2 +99 1 1 0.62514183353771979 0.5351855 0.2971879 0.16762662 1 2 0 +100 2 2 0.17792172249760466 0.83700794 0.15716855 0.005823519 2 1 0 +101 2 2 0.51406317914684485 0.5980606 0.36742735 0.03451207 2 1 0 +102 2 2 0.31702098316333388 0.7283155 0.25695667 0.014727903 2 1 0 +103 2 2 0.49735144806251907 0.6081392 0.3588682 0.032992546 2 1 0 +104 2 2 0.28794301589295634 0.7498043 0.23761658 0.01257914 2 1 0 +105 2 2 0.26518064209975606 0.7670674 0.2234128 0.009519864 2 1 0 +106 2 2 0.75950703348974569 0.46789703 0.4642631 0.06783989 2 1 0 +107 2 2 0.41810478246324223 0.65829325 0.31990713 0.021799628 2 1 0 +108 2 2 0.49937861526026439 0.60690767 0.36604086 0.027051508 2 1 0 +109 2 2 0.16413689798955344 0.84862584 0.1456836 0.0056905556 2 1 0 +110 2 2 0.42657364931760361 0.6527418 0.31644818 0.030810094 2 1 0 +111 2 2 0.48881501824463247 0.6133528 0.35600725 0.03063994 2 1 0 +112 2 2 0.3510193466525659 0.70397013 0.27720848 0.018821467 2 1 0 +113 2 2 0.48187621756246091 0.6176235 0.353735 0.028641492 2 1 0 +114 2 2 0.27751169051277835 0.7576667 0.23036094 0.011972419 2 1 0 +115 2 2 0.28013503173753401 0.7556817 0.23014143 0.014176883 2 1 0 +116 2 2 0.50263556149861677 0.6049342 0.3594847 0.03558109 2 1 0 +117 2 2 0.19968799985011623 0.81898624 0.17287412 0.008139691 2 1 0 +118 2 2 0.204429465821207 0.81511223 0.18027525 0.0046125324 2 1 0 +119 2 1 0.86035884522595385 0.50738645 0.42301026 0.06960334 1 2 0 +120 2 2 0.25214309728068479 0.7771335 0.21185966 0.011006939 2 1 0 +121 2 2 0.47448371210283841 0.6222062 0.34556687 0.032226894 2 1 0 +122 2 2 0.30199672781785708 0.7393405 0.2498451 0.010814363 2 1 0 +123 2 2 0.60323021952001576 0.5470417 0.40608206 0.046876296 2 1 0 +124 2 2 0.31923767548866339 0.7267028 0.2556534 0.017643875 2 1 0 +125 2 2 0.43175057493915137 0.6493713 0.32303485 0.027593875 2 1 0 +126 2 2 0.61069143612107968 0.5429753 0.40643388 0.050590873 2 1 0 +127 2 2 0.58245297106396454 0.55852664 0.39152822 0.049945217 2 1 0 +128 2 2 0.3531382372603053 0.7024801 0.2803944 0.017125545 2 1 0 +129 2 2 0.58665032833618669 0.5561872 0.39922902 0.04458387 2 1 0 +130 2 2 0.39546916952903632 0.67336404 0.30721402 0.019421963 2 1 0 +131 2 2 0.27882521851885139 0.75667214 0.22774182 0.015586088 2 1 0 +132 2 2 0.31255005927106427 0.731579 0.25469235 0.01372864 2 1 0 +133 2 1 0.79074193183673891 0.47120857 0.4535082 0.07528323 1 2 0 +134 2 1 0.80177714433697078 0.48472878 0.44853115 0.06674015 1 2 0 +135 2 2 0.23427676959660296 0.7911428 0.20064731 0.008209965 2 1 0 +136 2 2 0.22113523876673646 0.80160826 0.1887133 0.009678398 2 1 0 +137 2 2 0.49636329431544951 0.60874045 0.35492796 0.036331665 2 1 0 +138 2 2 0.59697535886568454 0.5504741 0.39672276 0.052803207 2 1 0 +139 2 2 0.35463379830153935 0.70143026 0.27821934 0.020350434 2 1 0 +140 2 2 0.2318067893087862 0.79309934 0.19774233 0.009158426 2 1 0 +141 2 2 0.29896378169244919 0.74158627 0.24273157 0.015682183 2 1 0 +142 2 2 0.51406317914684485 0.5980606 0.36742735 0.03451207 2 1 0 +143 2 2 0.23952097105524112 0.78700477 0.2032769 0.009718395 2 1 0 +144 2 2 0.19222648233640668 0.82512 0.16784589 0.007034164 2 1 0 +145 2 2 0.29608986432489648 0.7437206 0.2417711 0.014508302 2 1 0 +146 2 2 0.5393575997694513 0.58312273 0.3819506 0.034926824 2 1 0 +147 2 2 0.42746774173068397 0.65215844 0.31988582 0.02795579 2 1 0 +148 2 2 0.26484913163069684 0.7673217 0.21884738 0.013831014 2 1 0 +149 2 2 0.55522162684312593 0.57394505 0.38127565 0.044779327 2 1 0 diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris-out.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris-out.txt new file mode 100644 index 0000000000..7eab1bd9d9 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris-out.txt @@ -0,0 +1,151 @@ +maml.exe TrainTest test=%Data% tr=WeightedEnsembleMulticlass{bp=mlr{t-} nm=20 st=BootstrapSelector{} tp=-} dout=%Output% data=%Data% out=%Output% seed=1 xf=Term{col=Label} +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Training 20 learners for the batch 1 +Beginning training model 1 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 12 of 15 weights. +Trainer 1 of 20 finished in %Time% +Beginning training model 2 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 10 of 15 weights. +Trainer 2 of 20 finished in %Time% +Beginning training model 3 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 3 of 20 finished in %Time% +Beginning training model 4 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 10 of 15 weights. +Trainer 4 of 20 finished in %Time% +Beginning training model 5 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 5 of 20 finished in %Time% +Beginning training model 6 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 6 of 20 finished in %Time% +Beginning training model 7 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 7 of 20 finished in %Time% +Beginning training model 8 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 8 of 20 finished in %Time% +Beginning training model 9 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 9 of 20 finished in %Time% +Beginning training model 10 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 10 of 20 finished in %Time% +Beginning training model 11 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 12 of 15 weights. +Trainer 11 of 20 finished in %Time% +Beginning training model 12 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 12 of 20 finished in %Time% +Beginning training model 13 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 13 of 20 finished in %Time% +Beginning training model 14 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 12 of 15 weights. +Trainer 14 of 20 finished in %Time% +Beginning training model 15 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 15 of 20 finished in %Time% +Beginning training model 16 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 16 of 20 finished in %Time% +Beginning training model 17 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 10 of 15 weights. +Trainer 17 of 20 finished in %Time% +Beginning training model 18 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 18 of 20 finished in %Time% +Beginning training model 19 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 12 of 15 weights. +Trainer 19 of 20 finished in %Time% +Beginning training model 20 of 20 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 12 of 15 weights. +Trainer 20 of 20 finished in %Time% +Not training a calibrator because it is not needed. + +Confusion table + ||======================== +PREDICTED || 0 | 1 | 2 | Recall +TRUTH ||======================== + 0 || 50 | 0 | 0 | 1.0000 + 1 || 0 | 45 | 5 | 0.9000 + 2 || 0 | 4 | 46 | 0.9200 + ||======================== +Precision ||1.0000 |0.9184 |0.9020 | +Accuracy(micro-avg): 0.940000 +Accuracy(macro-avg): 0.940000 +Log-loss: 0.434802 +Log-loss reduction: 0.604226 + +OVERALL RESULTS +--------------------------------------- +Accuracy(micro-avg): 0.940000 (0.0000) +Accuracy(macro-avg): 0.940000 (0.0000) +Log-loss: 0.434802 (0.0000) +Log-loss reduction: 0.604226 (0.0000) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris-rp.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris-rp.txt new file mode 100644 index 0000000000..6d11d98efc --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris-rp.txt @@ -0,0 +1,4 @@ +WeightedEnsembleMulticlass +Accuracy(micro-avg) Accuracy(macro-avg) Log-loss Log-loss reduction /bp /nm Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings +0.94 0.94 0.434802 0.604226 mlr{t-} 20 WeightedEnsembleMulticlass %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=WeightedEnsembleMulticlass{bp=mlr{t-} nm=20 st=BootstrapSelector{} tp=-} dout=%Output% data=%Data% out=%Output% seed=1 xf=Term{col=Label} /bp:mlr{t-};/nm:20 + diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris.txt new file mode 100644 index 0000000000..a1f2b159ae --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Bootstrap-TrainTest-iris.txt @@ -0,0 +1,151 @@ +Instance Label Assigned Log-loss #1 Score #2 Score #3 Score #1 Class #2 Class #3 Class +0 0 0 0.19146156260105612 0.82575136 0.16329004 0.011343887 0 1 2 +1 0 0 0.21096344997532901 0.80980366 0.17785987 0.011758441 0 1 2 +2 0 0 0.19397905539384586 0.82367516 0.16559267 0.010766417 0 1 2 +3 0 0 0.21430201866978357 0.8071046 0.18038179 0.012308234 0 1 2 +4 0 0 0.18774933985517447 0.82882243 0.16011198 0.011200568 0 1 2 +5 0 0 0.25633429076259256 0.7738832 0.20344394 0.021169476 0 1 2 +6 0 0 0.22011507004015399 0.80242646 0.18346035 0.014116207 0 1 2 +7 0 0 0.20283311738421092 0.8164145 0.17206511 0.012108254 0 1 2 +8 0 0 0.2148520567453098 0.8066608 0.18086928 0.011703907 0 1 2 +9 0 0 0.18928539772866212 0.8275503 0.16312541 0.009787849 0 1 2 +10 0 0 0.19243678795545921 0.82494646 0.1631738 0.011981992 0 1 2 +11 0 0 0.21069858489986876 0.8100182 0.17774996 0.01279643 0 1 2 +12 0 0 0.18608395679206061 0.8302039 0.16020055 0.009295385 0 1 2 +13 0 0 0.16378879545906469 0.8489213 0.14343704 0.007628629 0 1 2 +14 0 0 0.15989203684182007 0.8522358 0.13795774 0.009677928 0 1 2 +15 0 0 0.20989580783965758 0.8106687 0.16973507 0.018127272 0 1 2 +16 0 0 0.21820461872064956 0.8039609 0.17832217 0.016521387 0 1 2 +17 0 0 0.21700444000945679 0.8049264 0.18108252 0.014253864 0 1 2 +18 0 0 0.23211459259808195 0.79285526 0.18940419 0.01715782 0 1 2 +19 0 0 0.21258898598306464 0.80848837 0.17653638 0.014837928 0 1 2 +20 0 0 0.22031014968058699 0.80226994 0.18441546 0.013839974 0 1 2 +21 0 0 0.24610642523150209 0.781839 0.19905026 0.01870889 0 1 2 +22 0 0 0.15897094297361211 0.85302114 0.1377921 0.00855796 0 1 2 +23 0 0 0.32366789594363532 0.7234905 0.24987638 0.027012125 0 1 2 +24 0 0 0.2372031818920764 0.788831 0.19627658 0.015387548 0 1 2 +25 0 0 0.22851782516239214 0.7957121 0.19077209 0.013283196 0 1 2 +26 0 0 0.27045379958967952 0.76303315 0.21731067 0.020228576 0 1 2 +27 0 0 0.19942101093228148 0.8192049 0.16928631 0.012094513 0 1 2 +28 0 0 0.19524145042317115 0.822636 0.16651651 0.011475574 0 1 2 +29 0 0 0.21846398964602987 0.8037524 0.18387958 0.013022162 0 1 2 +30 0 0 0.2231733391790903 0.79997617 0.18715349 0.01314823 0 1 2 +31 0 0 0.260962964032871 0.77030945 0.21112762 0.01942033 0 1 2 +32 0 0 0.15463442890268481 0.8567283 0.13386649 0.009110154 0 1 2 +33 0 0 0.16510593089613027 0.8478039 0.14033367 0.010777818 0 1 2 +34 0 0 0.18928539772866212 0.8275503 0.16312541 0.009787849 0 1 2 +35 0 0 0.18686632775198891 0.8295546 0.16040263 0.010183046 0 1 2 +36 0 0 0.18457000952733715 0.8314617 0.15822938 0.010771405 0 1 2 +37 0 0 0.18928539772866212 0.8275503 0.16312541 0.009787849 0 1 2 +38 0 0 0.20171934642363334 0.8173243 0.17116725 0.010882555 0 1 2 +39 0 0 0.20300192572187448 0.81627667 0.17216635 0.012149481 0 1 2 +40 0 0 0.20837513836989668 0.8119024 0.17472193 0.0133031 0 1 2 +41 0 0 0.2713003038560251 0.7623875 0.22223103 0.014569156 0 1 2 +42 0 0 0.19325292662708521 0.82427347 0.16498834 0.010676604 0 1 2 +43 0 0 0.33914975386361185 0.71237576 0.25531232 0.031025369 0 1 2 +44 0 0 0.28251571796968816 0.7538848 0.22091596 0.023833651 0 1 2 +45 0 0 0.23868928058933403 0.7876596 0.19652337 0.014711941 0 1 2 +46 0 0 0.19531086536609146 0.8225789 0.16497064 0.012499476 0 1 2 +47 0 0 0.20167406005791733 0.8173613 0.17132336 0.011444029 0 1 2 +48 0 0 0.19222005321517471 0.8251253 0.16301689 0.01194074 0 1 2 +49 0 0 0.19859640789756661 0.8198807 0.16926005 0.011478923 0 1 2 +50 1 1 0.69366492129978408 0.4997412 0.390107 0.11885977 1 2 0 +51 1 1 0.722177114765165 0.4856937 0.41258383 0.11151006 1 2 0 +52 1 1 0.74613877892215708 0.474194 0.44307593 0.08920108 1 2 0 +53 1 1 0.58995705400069653 0.5543511 0.28060544 0.16264556 1 2 0 +54 1 1 0.69727526440932153 0.4979402 0.40902108 0.09775919 1 2 0 +55 1 1 0.63299771361825929 0.53099763 0.32113132 0.14302845 1 2 0 +56 1 2 0.80045949283787321 0.47099262 0.44912255 0.08745068 2 1 0 +57 1 1 0.64462844168563782 0.5248575 0.31864232 0.14911339 1 0 2 +58 1 1 0.63999987490349286 0.5272925 0.33323693 0.13883023 1 2 0 +59 1 1 0.64337310433745454 0.5255168 0.30958351 0.15970789 1 2 0 +60 1 1 0.57967598218780425 0.5600798 0.27849436 0.15828522 1 0 2 +61 1 1 0.694414940157003 0.49936652 0.38107106 0.12467271 1 2 0 +62 1 1 0.56022738446740428 0.5710792 0.237293 0.18947217 1 0 2 +63 1 1 0.6717778471153717 0.51079965 0.37781808 0.112865895 1 2 0 +64 1 1 0.6435556153894364 0.5254209 0.25743625 0.2163626 1 2 0 +65 1 1 0.67429191631892671 0.5095171 0.36260965 0.13451363 1 2 0 +66 1 1 0.70564803008987942 0.49378848 0.40332592 0.10823792 1 2 0 +67 1 1 0.59226669986140101 0.5530722 0.25057793 0.19768134 1 0 2 +68 1 1 0.65503094596991529 0.519426 0.38644782 0.09157355 1 2 0 +69 1 1 0.58726876314771581 0.55584335 0.23047645 0.21299398 1 0 2 +70 1 2 0.92691433006856649 0.5489316 0.39577305 0.057030033 2 1 0 +71 1 1 0.63108365843646563 0.53201497 0.2865383 0.17858174 1 2 0 +72 1 1 0.69630746006737787 0.49842235 0.4223566 0.07916129 1 2 0 +73 1 1 0.61660993434853617 0.5397712 0.30047223 0.15105256 1 2 0 +74 1 1 0.63704713684969339 0.52885175 0.3101462 0.16002357 1 2 0 +75 1 1 0.66757887371019631 0.512949 0.35957545 0.13214076 1 2 0 +76 1 1 0.6700877539491773 0.5116637 0.3855137 0.10485341 1 2 0 +77 1 2 0.87022725180836724 0.5232084 0.41885635 0.058479164 2 1 0 +78 1 1 0.69888170354403267 0.49714094 0.40212 0.10550495 1 2 0 +79 1 1 0.63935960063722164 0.5276302 0.30466634 0.16351411 1 0 2 +80 1 1 0.58531710154346184 0.55692923 0.23551382 0.20625782 1 0 2 +81 1 1 0.60114899827081492 0.5481814 0.27571127 0.17337778 1 0 2 +82 1 1 0.60765136640193129 0.5446285 0.24525051 0.21019095 1 2 0 +83 1 2 0.78191902849971295 0.48279703 0.45752716 0.06242057 2 1 0 +84 1 1 0.70614740660406139 0.49354196 0.40246436 0.108545415 1 2 0 +85 1 2 0.79261248535426554 0.45707804 0.45266068 0.09905711 2 1 0 +86 1 1 0.72683155649442366 0.4834383 0.42676643 0.098733425 1 2 0 +87 1 1 0.59528731118220901 0.5514041 0.3088326 0.13472809 1 2 0 +88 1 1 0.64889456916206567 0.5226232 0.29479077 0.17990224 1 2 0 +89 1 1 0.60544714020316892 0.5458303 0.2826715 0.1687907 1 2 0 +90 1 1 0.59582025894026491 0.5511103 0.2770579 0.16660616 1 2 0 +91 1 1 0.67398733995141735 0.5096723 0.37292713 0.12050183 1 2 0 +92 1 1 0.59706577570415509 0.55042434 0.2509837 0.19781126 1 2 0 +93 1 1 0.63285292121159364 0.5310745 0.3125384 0.14885376 1 0 2 +94 1 1 0.62449908997933723 0.5355296 0.29871812 0.16074705 1 2 0 +95 1 1 0.63101184616066952 0.5320532 0.26759028 0.19765595 1 2 0 +96 1 1 0.64018086680858421 0.52719706 0.30100563 0.16847885 1 2 0 +97 1 1 0.63777142096307105 0.52846885 0.3093633 0.16043931 1 2 0 +98 1 1 0.66513654696216473 0.5142033 0.32096457 0.15675701 1 0 2 +99 1 1 0.63208665630471916 0.5314816 0.2929299 0.17162229 1 2 0 +100 2 2 0.19704112882922087 0.82115686 0.17259409 0.0062305043 2 1 0 +101 2 2 0.52693323230772071 0.59041286 0.37368852 0.035918117 2 1 0 +102 2 2 0.33383353061365767 0.716173 0.26873347 0.014969881 2 1 0 +103 2 2 0.51461925515875862 0.59772813 0.3656274 0.033569418 2 1 0 +104 2 2 0.30548520946867297 0.7367658 0.24955836 0.013095798 2 1 0 +105 2 2 0.28511648566894698 0.75192666 0.23750389 0.0096648075 2 1 0 +106 2 1 0.77045559113440054 0.46480623 0.46280217 0.070732296 1 2 0 +107 2 2 0.43462758939724772 0.64750576 0.3306417 0.021727186 2 1 0 +108 2 2 0.50667836130762822 0.6024935 0.36743265 0.027475437 2 1 0 +109 2 2 0.18239990420812591 0.83326805 0.16108054 0.0059612817 2 1 0 +110 2 2 0.44689714103761835 0.6396097 0.3293798 0.031963766 2 1 0 +111 2 2 0.50039590479779572 0.6062906 0.36107463 0.03155027 2 1 0 +112 2 2 0.36973314085733122 0.6909187 0.28880513 0.019333225 2 1 0 +113 2 2 0.48967989123533129 0.61282253 0.35691023 0.030163828 2 1 0 +114 2 2 0.29132059353693862 0.74727607 0.23866206 0.013091091 2 1 0 +115 2 2 0.30148235179491101 0.7397209 0.24577336 0.014947569 2 1 0 +116 2 2 0.52120738264322819 0.59380317 0.3673874 0.036394477 2 1 0 +117 2 2 0.22223259643984405 0.8007291 0.18996698 0.008388192 2 1 0 +118 2 2 0.22393114933863967 0.79937017 0.19535415 0.0047267694 2 1 0 +119 2 1 0.86335681019874533 0.50353104 0.421744 0.07112156 1 2 0 +120 2 2 0.27297948916515219 0.7611084 0.22682562 0.011377523 2 1 0 +121 2 2 0.48875826766613062 0.6133876 0.35361618 0.033884987 2 1 0 +122 2 2 0.32183527782518362 0.7248176 0.263148 0.010936804 2 1 0 +123 2 2 0.61157484936023432 0.54249585 0.40999818 0.04793093 2 1 0 +124 2 2 0.34160421771038918 0.7106294 0.27067435 0.01820742 2 1 0 +125 2 2 0.45394821245354905 0.6351156 0.33698285 0.027506717 2 1 0 +126 2 2 0.61981232397732444 0.5380454 0.41073883 0.0517139 2 1 0 +127 2 2 0.59622808170030672 0.5508856 0.3988101 0.051250108 2 1 0 +128 2 2 0.36819391247769245 0.691983 0.2895178 0.017895032 2 1 0 +129 2 2 0.60395363626275234 0.5466461 0.40726396 0.044511866 2 1 0 +130 2 2 0.40893660042290109 0.66435635 0.31609666 0.019469509 2 1 0 +131 2 2 0.30282171637880623 0.7387308 0.24597739 0.015782166 2 1 0 +132 2 2 0.32877311216879734 0.7198063 0.2652231 0.014460641 2 1 0 +133 2 1 0.80303858664981986 0.4800099 0.4479657 0.07599264 1 2 0 +134 2 1 0.81046582506585807 0.4880017 0.4446509 0.067600146 1 2 0 +135 2 2 0.2503820503895593 0.7785033 0.21323243 0.008407921 2 1 0 +136 2 2 0.24176277026101017 0.78524244 0.20471969 0.010243092 2 1 0 +137 2 2 0.51756669093110008 0.59596896 0.36444086 0.037151672 2 1 0 +138 2 2 0.60983139638861505 0.5434425 0.40329337 0.05418866 2 1 0 +139 2 2 0.37541503429501638 0.6870041 0.29133344 0.02088682 2 1 0 +140 2 2 0.25105419420584407 0.7779802 0.2124823 0.009635512 2 1 0 +141 2 2 0.31814927969015872 0.7274942 0.25642586 0.01635435 2 1 0 +142 2 2 0.52693323230772071 0.59041286 0.37368852 0.035918117 2 1 0 +143 2 2 0.25887062206618422 0.7719229 0.21752813 0.0100839175 2 1 0 +144 2 2 0.2115366936060161 0.8093396 0.18358684 0.0074414867 2 1 0 +145 2 2 0.31374428423664041 0.73070586 0.25430053 0.015252672 2 1 0 +146 2 2 0.54738550476984382 0.5784602 0.38503945 0.03611377 2 1 0 +147 2 2 0.4458575050344955 0.640275 0.3312786 0.028866546 2 1 0 +148 2 2 0.28699669501592695 0.7505142 0.2351964 0.0145819485 2 1 0 +149 2 2 0.57125843388083719 0.5648142 0.38956338 0.045926288 2 1 0 diff --git a/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Voting-TrainTest-iris-out.txt b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Voting-TrainTest-iris-out.txt new file mode 100644 index 0000000000..984626c121 --- /dev/null +++ b/test/BaselineOutput/Common/WeightedEnsembleMulticlass/net9.0/WE-Voting-TrainTest-iris-out.txt @@ -0,0 +1,61 @@ +maml.exe TrainTest test=%Data% tr=WeightedEnsembleMulticlass{bp=mlr{t-} nm=5 oc=MultiVoting tp=-} dout=%Output% data=%Data% out=%Output% seed=1 xf=Term{col=Label} +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Training 5 learners for the batch 1 +Beginning training model 1 of 5 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 12 of 15 weights. +Trainer 1 of 5 finished in %Time% +Beginning training model 2 of 5 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 10 of 15 weights. +Trainer 2 of 5 finished in %Time% +Beginning training model 3 of 5 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 3 of 5 finished in %Time% +Beginning training model 4 of 5 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 10 of 15 weights. +Trainer 4 of 5 finished in %Time% +Beginning training model 5 of 5 +Beginning optimization +num vars: 15 +improvement criterion: Mean Improvement +L1 regularization selected 11 of 15 weights. +Trainer 5 of 5 finished in %Time% +Not training a calibrator because it is not needed. + +Confusion table + ||======================== +PREDICTED || 0 | 1 | 2 | Recall +TRUTH ||======================== + 0 || 50 | 0 | 0 | 1.0000 + 1 || 0 | 44 | 6 | 0.8800 + 2 || 0 | 2 | 48 | 0.9600 + ||======================== +Precision ||1.0000 |0.9565 |0.8889 | +Accuracy(micro-avg): 0.946667 +Accuracy(macro-avg): 0.946667 +Log-loss: 0.511576 +Log-loss reduction: 0.534344 + +OVERALL RESULTS +--------------------------------------- +Accuracy(micro-avg): 0.946667 (0.0000) +Accuracy(macro-avg): 0.946667 (0.0000) +Log-loss: 0.511576 (0.0000) +Log-loss reduction: 0.534344 (0.0000) + +--------------------------------------- +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + From 733c9e39c67324e225f2f57dd466e7449a324cae Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Wed, 27 Mar 2024 13:19:30 -0700 Subject: [PATCH 18/26] Update baselines for Microsoft.ML.TestFramework on net8.0 --- .../CommandTrainMlrWithStats-out.txt | 4 +-- .../CommandTrainMlrWithStats-summary.txt | 26 +++++++++---------- .../CommandTrainingLrWithStats-out.txt | 17 ++++++++++++ .../CommandTrainingLrWithStats-summary.txt | 15 +++++++++++ .../TestParquetPrimitiveDataTypes-Data.txt | 15 +++++++++++ 5 files changed, 62 insertions(+), 15 deletions(-) create mode 100644 test/BaselineOutput/Common/Command/netcoreapp/CommandTrainingLrWithStats-out.txt create mode 100644 test/BaselineOutput/Common/Command/netcoreapp/CommandTrainingLrWithStats-summary.txt create mode 100644 test/BaselineOutput/Common/SavePipe/netcoreapp/TestParquetPrimitiveDataTypes-Data.txt diff --git a/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainMlrWithStats-out.txt b/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainMlrWithStats-out.txt index 15ca05f14a..7192a04a0b 100644 --- a/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainMlrWithStats-out.txt +++ b/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainMlrWithStats-out.txt @@ -5,9 +5,9 @@ num vars: 15 improvement criterion: Mean Improvement L1 regularization selected 11 of 15 weights. Model trained with 150 training examples. -Residual Deviance: 132.25266 +Residual Deviance: 132.04128 Null Deviance: 329.58368 -AIC: 154.25266 +AIC: 154.04128 Not training a calibrator because it is not needed. Physical memory usage(MB): %Number% Virtual memory usage(MB): %Number% diff --git a/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainMlrWithStats-summary.txt b/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainMlrWithStats-summary.txt index 673f8d8ba6..10ddc21eb6 100644 --- a/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainMlrWithStats-summary.txt +++ b/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainMlrWithStats-summary.txt @@ -1,18 +1,18 @@ LbfgsMaximumEntropyMulticlassTrainer bias and non-zero weights - Iris-setosa+(Bias) 2.2461631 - Iris-versicolor+(Bias) 0.77507037 - Iris-virginica+(Bias) -3.0212474 - Iris-setosa+f3 -3.1763735 - Iris-setosa+f2 -2.8875842 - Iris-setosa+f1 0.51886666 - Iris-setosa+f0 -0.0041847904 - Iris-versicolor+f1 -0.73153543 - Iris-virginica+f3 3.147383 - Iris-virginica+f2 1.9020821 - Iris-virginica+f0 0.0012227822 + Iris-setosa+(Bias) 2.293142 + Iris-versicolor+(Bias) 0.76599276 + Iris-virginica+(Bias) -3.0591424 + Iris-setosa+f3 -3.1762753 + Iris-setosa+f2 -2.8880758 + Iris-setosa+f1 0.5144032 + Iris-setosa+f0 -0.05904869 + Iris-versicolor+f1 -0.70699453 + Iris-virginica+f3 3.154204 + Iris-virginica+f2 1.9121273 + Iris-virginica+f0 0.036154807 *** MODEL STATISTICS SUMMARY *** Count of training examples: 150 -Residual Deviance: 132.25266 +Residual Deviance: 132.04128 Null Deviance: 329.58368 -AIC: 154.25266 +AIC: 154.04128 diff --git a/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainingLrWithStats-out.txt b/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainingLrWithStats-out.txt new file mode 100644 index 0000000000..d735624ab3 --- /dev/null +++ b/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainingLrWithStats-out.txt @@ -0,0 +1,17 @@ +maml.exe Train feat=Num lab=Lab tr=lr{t=- stat=+} loader=text{header+ col=Lab:0 col=Num:9-14} data=%Data% out=%Output% +Automatically adding a MinMax normalization transform, use 'norm=Warn' or 'norm=No' to turn this behavior off. +Beginning optimization +num vars: 7 +improvement criterion: Mean Improvement +Warning: Premature convergence occurred. The OptimizationTolerance may be set too small. ro equals zero. Is your function linear? +L1 regularization selected 7 of 7 weights. +Model trained with 500 training examples. +Residual Deviance: 458.97424 (on 493 degrees of freedom) +Null Deviance: 539.27637 (on 499 degrees of freedom) +AIC: 472.97424 +Warning: The number of parameters is too large. Cannot hold the variance-covariance matrix in memory. Skipping computation of standard errors and z-statistics of coefficients. Consider choosing a larger L1 regularizerto reduce the number of parameters. +Not training a calibrator because it is not needed. +Physical memory usage(MB): %Number% +Virtual memory usage(MB): %Number% +%DateTime% Time elapsed(s): %Number% + diff --git a/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainingLrWithStats-summary.txt b/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainingLrWithStats-summary.txt new file mode 100644 index 0000000000..8ae92790ce --- /dev/null +++ b/test/BaselineOutput/Common/Command/netcoreapp/CommandTrainingLrWithStats-summary.txt @@ -0,0 +1,15 @@ +Linear Binary Classification Predictor non-zero weights + +(Bias) -4.4260798 +education-num 2.1022286 +age 1.9192384 +hours-per-week 1.8831699 +capital-gain 1.6710597 +capital-loss 0.97718734 +fnlwgt 0.3190605 + +*** MODEL STATISTICS SUMMARY *** +Count of training examples: 500 +Residual Deviance: 458.97424 +Null Deviance: 539.27637 +AIC: 472.97424 diff --git a/test/BaselineOutput/Common/SavePipe/netcoreapp/TestParquetPrimitiveDataTypes-Data.txt b/test/BaselineOutput/Common/SavePipe/netcoreapp/TestParquetPrimitiveDataTypes-Data.txt new file mode 100644 index 0000000000..54ea2d51ed --- /dev/null +++ b/test/BaselineOutput/Common/SavePipe/netcoreapp/TestParquetPrimitiveDataTypes-Data.txt @@ -0,0 +1,15 @@ +#@ TextLoader{ +#@ header+ +#@ sep=tab +#@ col=sbyte:I1:0 +#@ col=short:I2:1 +#@ col=int:I4:2 +#@ col=long:I8:3 +#@ col=bool:BL:4 +#@ col=DateTimeOffset:DZ:5 +#@ col=Interval:TS:6 +#@ col=string:TX:7 +#@ } +sbyte short int long bool DateTimeOffset Interval string +-128 -32768 -2147483648 -9223372036854775808 1 "2018-09-01T19:53:18.2908167+00:00" "31.00:00:00.0010000" "" +127 32767 2147483647 9223372036854775807 0 "2018-09-01T19:53:18.3108175+00:00" "31.00:00:00.0010000" """""" From c48f24466f25896c77858f4e7eac6de3a8683ab7 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Wed, 27 Mar 2024 13:21:41 -0700 Subject: [PATCH 19/26] Update baselines for Microsoft.ML.TestFramework on net9.0 --- .../CommandTrainMlrWithStats-summary.txt | 18 ++++++++++++++++++ 1 file changed, 18 insertions(+) create mode 100644 test/BaselineOutput/Common/Command/net9.0/CommandTrainMlrWithStats-summary.txt diff --git a/test/BaselineOutput/Common/Command/net9.0/CommandTrainMlrWithStats-summary.txt b/test/BaselineOutput/Common/Command/net9.0/CommandTrainMlrWithStats-summary.txt new file mode 100644 index 0000000000..abec0ff738 --- /dev/null +++ b/test/BaselineOutput/Common/Command/net9.0/CommandTrainMlrWithStats-summary.txt @@ -0,0 +1,18 @@ +LbfgsMaximumEntropyMulticlassTrainer bias and non-zero weights + Iris-setosa+(Bias) 2.2931674 + Iris-versicolor+(Bias) 0.7660589 + Iris-virginica+(Bias) -3.059276 + Iris-setosa+f3 -3.1763687 + Iris-setosa+f2 -2.8882244 + Iris-setosa+f1 0.51452595 + Iris-setosa+f0 -0.05890327 + Iris-versicolor+f1 -0.7073206 + Iris-virginica+f3 3.1549096 + Iris-virginica+f2 1.91261 + Iris-virginica+f0 0.035835333 + +*** MODEL STATISTICS SUMMARY *** +Count of training examples: 150 +Residual Deviance: 132.03015 +Null Deviance: 329.58368 +AIC: 154.03015 From 287f4a9a2045047c3fe3acf4857878d17fa32099 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Wed, 27 Mar 2024 13:35:45 -0700 Subject: [PATCH 20/26] Update baselines for Microsoft.ML.Tests on net8.0 --- .../netcoreapp/SDCARegression.tsv | 17 +++++++++++++++++ .../TrainerEstimators/SdcaTests.cs | 2 +- 2 files changed, 18 insertions(+), 1 deletion(-) create mode 100644 test/BaselineOutput/Common/FeatureContribution/netcoreapp/SDCARegression.tsv diff --git a/test/BaselineOutput/Common/FeatureContribution/netcoreapp/SDCARegression.tsv b/test/BaselineOutput/Common/FeatureContribution/netcoreapp/SDCARegression.tsv new file mode 100644 index 0000000000..8ea2cba7b3 --- /dev/null +++ b/test/BaselineOutput/Common/FeatureContribution/netcoreapp/SDCARegression.tsv @@ -0,0 +1,17 @@ +#@ TextLoader{ +#@ sep=tab +#@ col=X1:R4:0 +#@ col=X2VBuffer:R4:1-4 +#@ col=X3Important:R4:5 +#@ col=Label:R4:6 +#@ col=Features:R4:7-12 +#@ col=Features:R4:13-18 +#@ col=FeatureContributions:R4:19-24 +#@ col=FeatureContributions:R4:25-30 +#@ col=FeatureContributions:R4:31-36 +#@ col=FeatureContributions:R4:37-42 +#@ } +950 757 692 720 297 7515 184479 950 757 692 720 297 7515 0.95669687 0.760804 0.7872582 0.75471693 0.29789367 0.7578661 0.06895071 0 0 0.058869552 0 1 18 11:1 17:141970.19 +459 961 0 659 274 2147 66495 459 961 0 659 274 2147 0.46223566 0.96582913 0 0.69077563 0.27482447 0.21651876 0 0.24562587 0 0.18859954 0 1 18 11:1 17:40560.21 +672 275 0 65 195 9818 208477 672 275 0 65 195 9818 0.6767372 0.2763819 0 0.06813417 0.19558676 0.99011695 0.037332788 0.015370657 0 0 0 1 18 11:1 17:185477.48 +186 301 0 681 526 1456 46093 186 301 0 681 526 1456 0.18731117 0.30251256 0 0.71383643 0.52758276 0.1468334 0 0 0 0.2873909 0.17624117 1 18 11:1 17:27506.135 diff --git a/test/Microsoft.ML.Tests/TrainerEstimators/SdcaTests.cs b/test/Microsoft.ML.Tests/TrainerEstimators/SdcaTests.cs index 159f341071..1ed1fdcfa9 100644 --- a/test/Microsoft.ML.Tests/TrainerEstimators/SdcaTests.cs +++ b/test/Microsoft.ML.Tests/TrainerEstimators/SdcaTests.cs @@ -83,7 +83,7 @@ public void SdcaLogisticRegression() // Check a few metrics to make sure the trained model is ok. Assert.InRange(metrics.AreaUnderRocCurve, 0.9, 1); - Assert.InRange(metrics.LogLoss, 0, 0.5); + Assert.InRange(metrics.LogLoss, 0, 0.51); var rawPrediction = mlContext.Data.CreateEnumerable(prediction, false); From 703cc3cb03b74a9b55f3a567ada023a131aa10b2 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Thu, 28 Mar 2024 18:04:42 -0700 Subject: [PATCH 21/26] Add missing baseline --- .../Common/EntryPoints/core_manifest.json | 30768 ++++++++++++++++ 1 file changed, 30768 insertions(+) create mode 100644 test/BaselineOutput/Common/EntryPoints/core_manifest.json diff --git a/test/BaselineOutput/Common/EntryPoints/core_manifest.json b/test/BaselineOutput/Common/EntryPoints/core_manifest.json new file mode 100644 index 0000000000..3dbaad5989 --- /dev/null +++ b/test/BaselineOutput/Common/EntryPoints/core_manifest.json @@ -0,0 +1,30768 @@ +{ + "EntryPoints": [ + { + "Name": "Data.CustomTextLoader", + "Desc": "Import a dataset from a text file", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "InputFile", + "Type": "FileHandle", + "Desc": "Location of the input file", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "CustomSchema", + "Type": "String", + "Desc": "Custom schema to use for parsing", + "Aliases": [ + "schema" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The resulting data view" + } + ] + }, + { + "Name": "Data.DataViewReference", + "Desc": "Pass dataview from memory to experiment", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Pointer to IDataView in memory", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The resulting data view" + } + ] + }, + { + "Name": "Data.IDataViewArrayConverter", + "Desc": "Create an array variable of IDataView", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "The data sets", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "The data set array" + } + ] + }, + { + "Name": "Data.PredictorModelArrayConverter", + "Desc": "Create an array variable of PredictorModel", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputModels", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The model array" + } + ] + }, + { + "Name": "Data.TextLoader", + "Desc": "Import a dataset from a text file", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "InputFile", + "Type": "FileHandle", + "Desc": "Location of the input file", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Arguments", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the column", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Type", + "Type": { + "Kind": "Enum", + "Values": [ + "I1", + "U1", + "I2", + "U2", + "I4", + "U4", + "I8", + "U8", + "R4", + "Num", + "R8", + "TX", + "Text", + "TXT", + "BL", + "Bool", + "TimeSpan", + "TS", + "DT", + "DateTime", + "DZ", + "DateTimeZone", + "UG", + "U16" + ] + }, + "Desc": "Type of the items in the column", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "R4" + }, + { + "Name": "Source", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Min", + "Type": "Int", + "Desc": "First index in the range", + "Required": true, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Max", + "Type": "Int", + "Desc": "Last index in the range", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "AutoEnd", + "Type": "Bool", + "Desc": "This range extends to the end of the line, but should be a fixed number of items", + "Aliases": [ + "auto" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "VariableEnd", + "Type": "Bool", + "Desc": "This range extends to the end of the line, which can vary from line to line", + "Aliases": [ + "var" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "AllOther", + "Type": "Bool", + "Desc": "This range includes only other indices not specified", + "Aliases": [ + "other" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "ForceVector", + "Type": "Bool", + "Desc": "Force scalar columns to be treated as vectors of length one", + "Aliases": [ + "vector" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + } + ] + } + }, + "Desc": "Source index range(s) of the column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "KeyCount", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Count", + "Type": "UInt", + "Desc": "Count of valid key values", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ] + }, + "Desc": "For a key column, this defines the range of values", + "Aliases": [ + "key" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "Column groups. Each group is specified as name:type:numeric-ranges, eg, col=Features:R4:1-17,26,35-40", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "AllowQuoting", + "Type": "Bool", + "Desc": "Whether the input may include double-quoted values. This parameter is used to distinguish separator characters in an input value from actual separators. When true, separators within double quotes are treated as part of the input value. When false, all separators, even those within quotes, are treated as delimiting a new column.", + "Aliases": [ + "quote" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "AllowSparse", + "Type": "Bool", + "Desc": "Whether the input may include sparse representations", + "Aliases": [ + "sparse" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "InputSize", + "Type": "Int", + "Desc": "Number of source columns in the text data. Default is that sparse rows contain their size information.", + "Aliases": [ + "size" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Separator", + "Type": { + "Kind": "Array", + "ItemType": "Char" + }, + "Desc": "Source column separator.", + "Aliases": [ + "sep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + "\t" + ] + }, + { + "Name": "Decimal Marker", + "Type": "Char", + "Desc": "Character symbol used to separate the integer part from the fractional part of a number written in decimal form.", + "Aliases": [ + "decimal" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "." + }, + { + "Name": "TrimWhitespace", + "Type": "Bool", + "Desc": "Remove trailing whitespace from lines", + "Aliases": [ + "trim" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "HasHeader", + "Type": "Bool", + "Desc": "Data file has header with feature names. Header is read only if options 'hs' and 'hf' are not specified.", + "Aliases": [ + "header" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseThreads", + "Type": "Bool", + "Desc": "Use separate parsing threads?", + "Aliases": [ + "threads" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "ReadMultilines", + "Type": "Bool", + "Desc": "Escape new line characters inside a quoted field? If AllowQuoting is false, this argument is ignored.", + "Aliases": [ + "multilines" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "HeaderFile", + "Type": "String", + "Desc": "File containing a header with feature names. If specified, header defined in the data file (header+) is ignored.", + "Aliases": [ + "hf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "MaxRows", + "Type": "Int", + "Desc": "Maximum number of rows to produce", + "Aliases": [ + "rows" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "EscapeChar", + "Type": "Char", + "Desc": "Character to use to escape quotes inside quoted fields. It can't be a character used as separator.", + "Aliases": [ + "escapechar" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "\"" + }, + { + "Name": "MissingRealsAsNaNs", + "Type": "Bool", + "Desc": "If true, empty float fields will be loaded as NaN. If false, they'll be loaded as 0. Default is false.", + "Aliases": [ + "missingrealnan" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + } + ] + }, + "Desc": "Arguments", + "Aliases": [ + "args" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The resulting data view" + } + ] + }, + { + "Name": "Models.AnomalyDetectionEvaluator", + "Desc": "Evaluates an anomaly detection scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "K", + "Type": "Int", + "Desc": "Expected number of false positives", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "P", + "Type": "Float", + "Desc": "Expected false positive rate", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01 + }, + { + "Name": "NumTopResults", + "Type": "Int", + "Desc": "Number of top-scored predictions to display", + "Aliases": [ + "n" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 50 + }, + { + "Name": "Stream", + "Type": "Bool", + "Desc": "Whether to calculate metrics in one pass", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxAucExamples", + "Type": "Int", + "Desc": "The number of samples to use for AUC calculation. If 0, AUC is not computed. If -1, the whole dataset is used", + "Aliases": [ + "numauc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.AnomalyPipelineEnsemble", + "Desc": "Combine anomaly detection models into an ensemble", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Average" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IAnomalyDetectionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.BinaryClassificationEvaluator", + "Desc": "Evaluates a binary classification scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "ProbabilityColumn", + "Type": "String", + "Desc": "Probability column name", + "Aliases": [ + "prob" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Threshold", + "Type": "Float", + "Desc": "Probability value for classification thresholding", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "UseRawScoreThreshold", + "Type": "Bool", + "Desc": "Use raw score value instead of probability for classification thresholding", + "Aliases": [ + "useRawScore" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumRocExamples", + "Type": "Int", + "Desc": "The number of samples to use for p/r curve generation. Specify 0 for no p/r curve generation", + "Aliases": [ + "numpr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100000 + }, + { + "Name": "MaxAucExamples", + "Type": "Int", + "Desc": "The number of samples to use for AUC calculation. If 0, AUC is not computed. If -1, the whole dataset is used", + "Aliases": [ + "numauc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "NumAuPrcExamples", + "Type": "Int", + "Desc": "The number of samples to use for AUPRC calculation. Specify 0 for no AUPRC calculation", + "Aliases": [ + "numauprc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100000 + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + }, + { + "Name": "ConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IClassificationEvaluatorOutput", + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.BinaryEnsemble", + "Desc": "Combine binary classifiers into an ensemble", + "FriendlyName": "Parallel Ensemble (bagging, stacking, etc)", + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average", + "Vote" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Median" + }, + { + "Name": "ValidatePipelines", + "Type": "Bool", + "Desc": "Whether to validate that all the pipelines are identical", + "Aliases": [ + "validate" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.BinaryPipelineEnsemble", + "Desc": "Combine binary classification models into an ensemble", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average", + "Vote" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Median" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.ClassificationEvaluator", + "Desc": "Evaluates a multi class classification scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "OutputTopKAcc", + "Type": "Int", + "Desc": "Output top-K accuracy.", + "Aliases": [ + "topkacc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumTopClassesToOutput", + "Type": "Int", + "Desc": "Output top-K classes.", + "Aliases": [ + "topk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 3 + }, + { + "Name": "NumClassesConfusionMatrix", + "Type": "Int", + "Desc": "Maximum number of classes in confusion matrix.", + "Aliases": [ + "nccf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "OutputPerClassStatistics", + "Type": "Bool", + "Desc": "Output per class statistics and confusion matrix.", + "Aliases": [ + "opcs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + }, + { + "Name": "ConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IClassificationEvaluatorOutput", + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.ClusterEvaluator", + "Desc": "Evaluates a clustering scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "FeatureColumn", + "Type": "String", + "Desc": "Features column name", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "CalculateDbi", + "Type": "Bool", + "Desc": "Calculate DBI? (time-consuming unsupervised metric)", + "Aliases": [ + "dbi" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NumTopClustersToOutput", + "Type": "Int", + "Desc": "Output top K clusters", + "Aliases": [ + "topk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 3 + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.CrossValidationResultsCombiner", + "Desc": "Combine the metric data views returned from cross validation.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "OverallMetrics", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Overall metrics datasets", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PerInstanceMetrics", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Per instance metrics datasets", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ConfusionMatrix", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Confusion matrix datasets", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Warnings", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Warning datasets", + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Kind", + "Type": { + "Kind": "Enum", + "Values": [ + "SignatureBinaryClassifierTrainer", + "SignatureMulticlassClassificationTrainer", + "SignatureRankerTrainer", + "SignatureRegressorTrainer", + "SignatureMultiOutputRegressorTrainer", + "SignatureAnomalyDetectorTrainer", + "SignatureClusteringTrainer" + ] + }, + "Desc": "Specifies the trainer kind, which determines the evaluator to be used.", + "Required": true, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "SignatureBinaryClassifierTrainer" + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "The label column name", + "Aliases": [ + "Label" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 7.0, + "IsNullable": false, + "Default": "Weight" + }, + { + "Name": "GroupColumn", + "Type": "String", + "Desc": "Column to use for grouping", + "Aliases": [ + "group" + ], + "Required": false, + "SortOrder": 8.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 9.0, + "IsNullable": false, + "Default": "Name" + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + }, + { + "Name": "ConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset" + } + ] + }, + { + "Name": "Models.CrossValidator", + "Desc": "Cross validation for general learning", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data set", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "TransformModel", + "Type": "TransformModel", + "Desc": "The transform model from the pipeline before this command. It gets included in the Output.PredictorModel.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Nodes", + "Type": { + "Kind": "Array", + "ItemType": "Node" + }, + "Desc": "The training subgraph", + "Required": true, + "SortOrder": 3.0, + "IsNullable": false + }, + { + "Name": "Inputs", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for training", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ] + }, + "Desc": "The training subgraph inputs", + "Required": true, + "SortOrder": 4.0, + "IsNullable": false + }, + { + "Name": "Outputs", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + } + ] + }, + "Desc": "The training subgraph outputs", + "Required": true, + "SortOrder": 5.0, + "IsNullable": false + }, + { + "Name": "StratificationColumn", + "Type": "String", + "Desc": "Column to use for stratification", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NumFolds", + "Type": "Int", + "Desc": "Number of folds in k-fold cross-validation", + "Aliases": [ + "k" + ], + "Required": false, + "SortOrder": 7.0, + "IsNullable": false, + "Default": 2 + }, + { + "Name": "Kind", + "Type": { + "Kind": "Enum", + "Values": [ + "SignatureBinaryClassifierTrainer", + "SignatureMulticlassClassificationTrainer", + "SignatureRankerTrainer", + "SignatureRegressorTrainer", + "SignatureMultiOutputRegressorTrainer", + "SignatureAnomalyDetectorTrainer", + "SignatureClusteringTrainer" + ] + }, + "Desc": "Specifies the trainer kind, which determines the evaluator to be used.", + "Required": true, + "SortOrder": 8.0, + "IsNullable": false, + "Default": "SignatureBinaryClassifierTrainer" + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 9.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 10.0, + "IsNullable": false, + "Default": "Weight" + }, + { + "Name": "GroupColumn", + "Type": "String", + "Desc": "Column to use for grouping", + "Aliases": [ + "group" + ], + "Required": false, + "SortOrder": 11.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 12.0, + "IsNullable": false, + "Default": "Name" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The final model including the trained predictor model and the model from the transforms, provided as the Input.TransformModel." + }, + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + }, + { + "Name": "ConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset" + } + ] + }, + { + "Name": "Models.CrossValidatorDatasetSplitter", + "Desc": "Split the dataset into the specified number of cross-validation folds (train and test sets)", + "FriendlyName": "Dataset CV Split", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumFolds", + "Type": "Int", + "Desc": "Number of folds to split into", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 2 + }, + { + "Name": "StratificationColumn", + "Type": "String", + "Desc": "Stratification column", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "TrainData", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Training data (one dataset per fold)" + }, + { + "Name": "TestData", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "Testing data (one dataset per fold)" + } + ] + }, + { + "Name": "Models.DatasetTransformer", + "Desc": "Applies a TransformModel to a dataset.", + "FriendlyName": "Apply Transform Model Output", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "TransformModel", + "Type": "TransformModel", + "Desc": "Transform model", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + } + ], + "InputKind": [ + "ITransformInput" + ] + }, + { + "Name": "Models.EnsembleSummary", + "Desc": "Summarize a pipeline ensemble predictor.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to summarize", + "Aliases": [ + "predictorModel" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "Summaries", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "The summaries of the individual predictors" + }, + { + "Name": "Stats", + "Type": { + "Kind": "Array", + "ItemType": "DataView" + }, + "Desc": "The model statistics of the individual predictors" + } + ] + }, + { + "Name": "Models.FixedPlattCalibrator", + "Desc": "Apply a Platt calibrator with a fixed slope and offset to an input model", + "FriendlyName": "Fixed Sigmoid Calibration", + "ShortName": null, + "Inputs": [ + { + "Name": "Slope", + "Type": "Float", + "Desc": "The slope parameter of the calibration function 1 / (1 + exp(slope * x + offset)", + "Aliases": [ + "slope" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": -1.0 + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UncalibratedPredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to calibrate", + "Aliases": [ + "uncalibratedPredictorModel" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "The offset parameter of the calibration function 1 / (1 + exp(slope * x + offset)", + "Aliases": [ + "offset" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "MaxRows", + "Type": "Int", + "Desc": "The maximum number of examples to train the calibrator on", + "Aliases": [ + "maxRows" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1000000000, + "Range": { + "Inf": 0, + "Max": 2147483647 + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ICalibratorInput", + "ITransformInput" + ], + "OutputKind": [ + "ICalibratorOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.MultiClassPipelineEnsemble", + "Desc": "Combine multiclass classifiers into an ensemble", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average", + "Vote" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Median" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.MultiOutputRegressionEvaluator", + "Desc": "Evaluates a multi output regression scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "RegressionLossFunction" + }, + "Desc": "Loss function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "SquaredLoss" + } + }, + { + "Name": "SuppressScoresAndLabels", + "Type": "Bool", + "Desc": "Suppress labels and scores in per-instance outputs?", + "Aliases": [ + "noScores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.NaiveCalibrator", + "Desc": "Apply a Naive calibrator to an input model", + "FriendlyName": "Naive Calibrator", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UncalibratedPredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to calibrate", + "Aliases": [ + "uncalibratedPredictorModel" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "MaxRows", + "Type": "Int", + "Desc": "The maximum number of examples to train the calibrator on", + "Aliases": [ + "maxRows" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1000000000, + "Range": { + "Inf": 0, + "Max": 2147483647 + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ICalibratorInput", + "ITransformInput" + ], + "OutputKind": [ + "ICalibratorOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.OneVersusAll", + "Desc": "One-vs-All macro (OVA)", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Nodes", + "Type": { + "Kind": "Array", + "ItemType": "Node" + }, + "Desc": "The subgraph for the binary trainer used to construct the OVA learner. This should be a TrainBinary node.", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "OutputForSubGraph", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Model", + "Type": "PredictorModel", + "Desc": "The predictor model for the subgraph exemplar.", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ] + }, + "Desc": "The training subgraph output.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "UseProbabilities", + "Type": "Bool", + "Desc": "Use probabilities in OVA combiner", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained multiclass model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ] + }, + { + "Name": "Models.OnnxConverter", + "Desc": "Converts the model to ONNX format.", + "FriendlyName": "ONNX Converter.", + "ShortName": null, + "Inputs": [ + { + "Name": "DataFile", + "Type": "String", + "Desc": "The data file", + "Aliases": [ + "data" + ], + "Required": false, + "SortOrder": 0.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Onnx", + "Type": "String", + "Desc": "The path to write the output ONNX to.", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Json", + "Type": "String", + "Desc": "The path to write the output JSON to.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The 'name' property in the output ONNX. By default this will be the ONNX extension-less name.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Domain", + "Type": "String", + "Desc": "The 'domain' property in the output ONNX.", + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "InputsToDrop", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Array of input column names to drop", + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "OutputsToDrop", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Array of output column names to drop", + "Required": false, + "SortOrder": 8.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Model that needs to be converted to ONNX format.", + "Required": false, + "SortOrder": 10.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "OnnxVersion", + "Type": { + "Kind": "Enum", + "Values": [ + "Stable", + "Experimental" + ] + }, + "Desc": "The targeted ONNX version. It can be either \"Stable\" or \"Experimental\". If \"Experimental\" is used, produced model can contain components that is not officially supported in ONNX standard.", + "Required": false, + "SortOrder": 11.0, + "IsNullable": false, + "Default": "Stable" + }, + { + "Name": "PredictiveModel", + "Type": "PredictorModel", + "Desc": "Predictor model that needs to be converted to ONNX format.", + "Required": false, + "SortOrder": 12.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [] + }, + { + "Name": "Models.OvaModelCombiner", + "Desc": "Combines a sequence of PredictorModels into a single model", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "ModelArray", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "Input models", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseProbabilities", + "Type": "Bool", + "Desc": "Use probabilities from learners instead of raw values.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Predictor model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ] + }, + { + "Name": "Models.PAVCalibrator", + "Desc": "Apply a PAV calibrator to an input model", + "FriendlyName": "PAV Calibration", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UncalibratedPredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to calibrate", + "Aliases": [ + "uncalibratedPredictorModel" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "MaxRows", + "Type": "Int", + "Desc": "The maximum number of examples to train the calibrator on", + "Aliases": [ + "maxRows" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1000000000, + "Range": { + "Inf": 0, + "Max": 2147483647 + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ICalibratorInput", + "ITransformInput" + ], + "OutputKind": [ + "ICalibratorOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.PlattCalibrator", + "Desc": "Apply a Platt calibrator to an input model", + "FriendlyName": "Sigmoid Calibration", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UncalibratedPredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to calibrate", + "Aliases": [ + "uncalibratedPredictorModel" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "MaxRows", + "Type": "Int", + "Desc": "The maximum number of examples to train the calibrator on", + "Aliases": [ + "maxRows" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1000000000, + "Range": { + "Inf": 0, + "Max": 2147483647 + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ICalibratorInput", + "ITransformInput" + ], + "OutputKind": [ + "ICalibratorOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.QuantileRegressionEvaluator", + "Desc": "Evaluates a quantile regression scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "RegressionLossFunction" + }, + "Desc": "Loss function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "SquaredLoss" + } + }, + { + "Name": "Index", + "Type": "Int", + "Desc": "Quantile index to select", + "Aliases": [ + "ind" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.RankingEvaluator", + "Desc": "Evaluates a ranking scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "GroupIdColumn", + "Type": "String", + "Desc": "Column to use for the group ID", + "Aliases": [ + "group" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "DcgTruncationLevel", + "Type": "Int", + "Desc": "Maximum truncation level for computing (N)DCG", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "LabelGains", + "Type": "String", + "Desc": "Label relevance gains", + "Aliases": [ + "gains" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "0,3,7,15,31" + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.RegressionEnsemble", + "Desc": "Combine regression models into an ensemble", + "FriendlyName": "Regression Ensemble (bagging, stacking, etc)", + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Median" + }, + { + "Name": "ValidatePipelines", + "Type": "Bool", + "Desc": "Whether to validate that all the pipelines are identical", + "Aliases": [ + "validate" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.RegressionEvaluator", + "Desc": "Evaluates a regression scored dataset.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for evaluation.", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name.", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Name" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "RegressionLossFunction" + }, + "Desc": "Loss function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "SquaredLoss" + } + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels.", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Weight column name.", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ScoreColumn", + "Type": "String", + "Desc": "Score column name.", + "Aliases": [ + "score" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "StratColumn", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Stratification column name.", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + } + ], + "InputKind": [ + "IEvaluatorInput" + ], + "OutputKind": [ + "IEvaluatorOutput" + ] + }, + { + "Name": "Models.RegressionPipelineEnsemble", + "Desc": "Combine regression models into an ensemble", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "PredictorModel" + }, + "Desc": "The models to combine into an ensemble", + "Aliases": [ + "models" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ModelCombiner", + "Type": { + "Kind": "Enum", + "Values": [ + "Median", + "Average" + ] + }, + "Desc": "The combiner used to combine the scores", + "Aliases": [ + "combiner" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Median" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Models.Summarizer", + "Desc": "Summarize a linear regression predictor.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor to summarize", + "Aliases": [ + "predictorModel" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "Summary", + "Type": "DataView", + "Desc": "The summary of a predictor" + }, + { + "Name": "Stats", + "Type": "DataView", + "Desc": "The training set statistics. Note that this output can be null." + } + ] + }, + { + "Name": "Models.TrainTestEvaluator", + "Desc": "General train test for any supported evaluator", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "train" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "TestingData", + "Type": "DataView", + "Desc": "The data to be used for testing", + "Aliases": [ + "test" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "TransformModel", + "Type": "TransformModel", + "Desc": "The aggregated transform model from the pipeline before this command, to apply to the test data, and also include in the final model, together with the predictor model.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Nodes", + "Type": { + "Kind": "Array", + "ItemType": "Node" + }, + "Desc": "The training subgraph", + "Required": true, + "SortOrder": 4.0, + "IsNullable": false + }, + { + "Name": "Inputs", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The data to be used for training", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ] + }, + "Desc": "The training subgraph inputs", + "Required": true, + "SortOrder": 5.0, + "IsNullable": false + }, + { + "Name": "Outputs", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + } + ] + }, + "Desc": "The training subgraph outputs", + "Required": true, + "SortOrder": 6.0, + "IsNullable": false + }, + { + "Name": "Kind", + "Type": { + "Kind": "Enum", + "Values": [ + "SignatureBinaryClassifierTrainer", + "SignatureMulticlassClassificationTrainer", + "SignatureRankerTrainer", + "SignatureRegressorTrainer", + "SignatureMultiOutputRegressorTrainer", + "SignatureAnomalyDetectorTrainer", + "SignatureClusteringTrainer" + ] + }, + "Desc": "Specifies the trainer kind, which determines the evaluator to be used.", + "Required": false, + "SortOrder": 7.0, + "IsNullable": false, + "Default": "SignatureBinaryClassifierTrainer" + }, + { + "Name": "PipelineId", + "Type": "String", + "Desc": "Identifies which pipeline was run for this train test.", + "Required": false, + "SortOrder": 8.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "IncludeTrainingMetrics", + "Type": "Bool", + "Desc": "Indicates whether to include and output training dataset metrics.", + "Required": false, + "SortOrder": 9.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 10.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 11.0, + "IsNullable": false, + "Default": "Weight" + }, + { + "Name": "GroupColumn", + "Type": "String", + "Desc": "Column to use for grouping", + "Aliases": [ + "group" + ], + "Required": false, + "SortOrder": 12.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NameColumn", + "Type": "String", + "Desc": "Name column name", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 13.0, + "IsNullable": false, + "Default": "Name" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The final model including the trained predictor model and the model from the transforms, provided as the Input.TransformModel." + }, + { + "Name": "Warnings", + "Type": "DataView", + "Desc": "Warning dataset" + }, + { + "Name": "OverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset" + }, + { + "Name": "ConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset" + }, + { + "Name": "TrainingWarnings", + "Type": "DataView", + "Desc": "Warning dataset for training" + }, + { + "Name": "TrainingOverallMetrics", + "Type": "DataView", + "Desc": "Overall metrics dataset for training" + }, + { + "Name": "TrainingPerInstanceMetrics", + "Type": "DataView", + "Desc": "Per instance metrics dataset for training" + }, + { + "Name": "TrainingConfusionMatrix", + "Type": "DataView", + "Desc": "Confusion matrix dataset for training" + } + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.ExponentialAverage", + "Desc": "Applies a Exponential average on a time series.", + "FriendlyName": "Exponential Average Transform", + "ShortName": "ExpAvg", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column", + "Aliases": [ + "name" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Decay", + "Type": "Float", + "Desc": "Coefficient d in: d m(y_t) = d * y_t + (1-d) * m(y_(t-1)), it should be in [0, 1].", + "Aliases": [ + "d" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.9 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.IidChangePointDetector", + "Desc": "This transform detects the change-points in an i.i.d. sequence using adaptive kernel density estimation and martingales.", + "FriendlyName": "IID Change Point Detection", + "ShortName": "ichgpnt", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column.", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Confidence", + "Type": "Float", + "Desc": "The confidence for change point detection in the range [0, 100].", + "Aliases": [ + "cnf" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 95.0 + }, + { + "Name": "ChangeHistoryLength", + "Type": "Int", + "Desc": "The length of the sliding window on p-values for computing the martingale score.", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": 20 + }, + { + "Name": "Martingale", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Power", + "Mixture" + ] + }, + "Desc": "The martingale used for scoring.", + "Aliases": [ + "mart" + ], + "Required": false, + "SortOrder": 103.0, + "IsNullable": false, + "Default": "Power" + }, + { + "Name": "PowerMartingaleEpsilon", + "Type": "Float", + "Desc": "The epsilon parameter for the Power martingale.", + "Aliases": [ + "eps" + ], + "Required": false, + "SortOrder": 104.0, + "IsNullable": false, + "Default": 0.1 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.IidSpikeDetector", + "Desc": "This transform detects the spikes in a i.i.d. sequence using adaptive kernel density estimation.", + "FriendlyName": "IID Spike Detection", + "ShortName": "ispike", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column.", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Confidence", + "Type": "Float", + "Desc": "The confidence for spike detection in the range [0, 100].", + "Aliases": [ + "cnf" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 99.0 + }, + { + "Name": "Side", + "Type": { + "Kind": "Enum", + "Values": [ + "Positive", + "Negative", + "TwoSided" + ] + }, + "Desc": "The argument that determines whether to detect positive or negative anomalies, or both.", + "Aliases": [ + "side" + ], + "Required": false, + "SortOrder": 101.0, + "IsNullable": false, + "Default": "TwoSided" + }, + { + "Name": "PvalueHistoryLength", + "Type": "Int", + "Desc": "The size of the sliding window for computing the p-value.", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": 100 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.PercentileThresholdTransform", + "Desc": "Detects the values of time-series that are in the top percentile of the sliding window.", + "FriendlyName": "Percentile Threshold Transform", + "ShortName": "TopPcnt", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column", + "Aliases": [ + "name" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Percentile", + "Type": "Float", + "Desc": "The percentile value for thresholding in the range [0, 100]", + "Aliases": [ + "pcnt" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The size of the sliding window for computing the percentile threshold. The default value is set to 1.", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 1 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.PValueTransform", + "Desc": "This P-Value transform calculates the p-value of the current input in the sequence with regard to the values in the sliding window.", + "FriendlyName": "p-Value Transform", + "ShortName": "PVal", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column", + "Aliases": [ + "name" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed value of the random generator", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "PositiveSide", + "Type": "Bool", + "Desc": "The flag that determines whether the p-values are calculated on the positive side", + "Aliases": [ + "pos" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The size of the sliding window for computing the p-value", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "InitialWindowSize", + "Type": "Int", + "Desc": "The size of the initial window for computing the p-value. The default value is set to 0, which means there is no initial window considered.", + "Aliases": [ + "initwnd" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.SlidingWindowTransform", + "Desc": "Returns the last values for a time series [y(t-d-l+1), y(t-d-l+2), ..., y(t-l-1), y(t-l)] where d is the size of the window, l the lag and y is a Float.", + "FriendlyName": "Sliding Window Transform", + "ShortName": "SlideWin", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The size of the sliding window for computing the moving average", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 2 + }, + { + "Name": "Lag", + "Type": "Int", + "Desc": "Lag between current observation and last observation from the sliding window", + "Aliases": [ + "l" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "Begin", + "Type": { + "Kind": "Enum", + "Values": [ + "NaNValues", + "FirstValue" + ] + }, + "Desc": "Define how to populate the first rows of the produced series", + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "NaNValues" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.SsaChangePointDetector", + "Desc": "This transform detects the change-points in a seasonal time-series using Singular Spectrum Analysis (SSA).", + "FriendlyName": "SSA Change Point Detection", + "ShortName": "chgpnt", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column.", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "TrainingWindowSize", + "Type": "Int", + "Desc": "The number of points from the beginning of the sequence used for training.", + "Aliases": [ + "twnd" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Confidence", + "Type": "Float", + "Desc": "The confidence for change point detection in the range [0, 100].", + "Aliases": [ + "cnf" + ], + "Required": true, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 95.0 + }, + { + "Name": "SeasonalWindowSize", + "Type": "Int", + "Desc": "An upper bound on the largest relevant seasonality in the input time-series.", + "Aliases": [ + "swnd" + ], + "Required": true, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "ChangeHistoryLength", + "Type": "Int", + "Desc": "The length of the sliding window on p-values for computing the martingale score.", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": 20 + }, + { + "Name": "ErrorFunction", + "Type": { + "Kind": "Enum", + "Values": [ + "SignedDifference", + "AbsoluteDifference", + "SignedProportion", + "AbsoluteProportion", + "SquaredDifference" + ] + }, + "Desc": "The function used to compute the error between the expected and the observed value.", + "Aliases": [ + "err" + ], + "Required": false, + "SortOrder": 103.0, + "IsNullable": false, + "Default": "SignedDifference" + }, + { + "Name": "Martingale", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Power", + "Mixture" + ] + }, + "Desc": "The martingale used for scoring.", + "Aliases": [ + "mart" + ], + "Required": false, + "SortOrder": 104.0, + "IsNullable": false, + "Default": "Power" + }, + { + "Name": "PowerMartingaleEpsilon", + "Type": "Float", + "Desc": "The epsilon parameter for the Power martingale.", + "Aliases": [ + "eps" + ], + "Required": false, + "SortOrder": 105.0, + "IsNullable": false, + "Default": 0.1 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.SsaForecasting", + "Desc": "This transform forecasts using Singular Spectrum Analysis (SSA).", + "FriendlyName": "SSA Forecasting", + "ShortName": "ssafcst", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column.", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The length of the window on the series for building the trajectory matrix (parameter L).", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "SeriesLength", + "Type": "Int", + "Desc": "The length of series that is kept in buffer for modeling (parameter N).", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "TrainSize", + "Type": "Int", + "Desc": "The length of series from the beginning used for training.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Horizon", + "Type": "Int", + "Desc": "The number of values to forecast.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "ConfidenceLevel", + "Type": "Float", + "Desc": "The confidence level in [0, 1) for forecasting.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0.95 + }, + { + "Name": "VariableHorizon", + "Type": "Bool", + "Desc": "Set this to true horizon will change at prediction time.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "ConfidenceLowerBoundColumn", + "Type": "String", + "Desc": "The name of the confidence interval lower bound column.", + "Aliases": [ + "cnfminname" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ConfidenceUpperBoundColumn", + "Type": "String", + "Desc": "The name of the confidence interval upper bound column.", + "Aliases": [ + "cnfmaxnname" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RankSelectionMethod", + "Type": { + "Kind": "Enum", + "Values": [ + "Fixed", + "Exact", + "Fast" + ] + }, + "Desc": "The rank selection method.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Exact" + }, + { + "Name": "Rank", + "Type": "Int", + "Desc": "The desired rank of the subspace used for SSA projection (parameter r). This parameter should be in the range in [1, windowSize]. If set to null, the rank is automatically determined based on prediction error minimization.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxRank", + "Type": "Int", + "Desc": "The maximum rank considered during the rank selection process. If not provided (i.e. set to null), it is set to windowSize - 1.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ShouldStabilize", + "Type": "Bool", + "Desc": "The flag determining whether the model should be stabilized.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "ShouldMaintainInfo", + "Type": "Bool", + "Desc": "The flag determining whether the meta information for the model needs to be maintained.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaxGrowth", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "TimeSpan", + "Type": "Int", + "Desc": "Time span of growth ratio. Must be strictly positive.", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Growth", + "Type": "Float", + "Desc": "Growth. Must be non-negative.", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0.0 + } + ] + }, + "Desc": "The maximum growth on the exponential trend.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DiscountFactor", + "Type": "Float", + "Desc": "The discount factor in [0,1] used for online updates.", + "Aliases": [ + "disc" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "IsAdaptive", + "Type": "Bool", + "Desc": "The flag determing whether the model is adaptive", + "Aliases": [ + "adp" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "TimeSeriesProcessingEntryPoints.SsaSpikeDetector", + "Desc": "This transform detects the spikes in a seasonal time-series using Singular Spectrum Analysis (SSA).", + "FriendlyName": "SSA Spike Detection", + "ShortName": "spike", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "The name of the source column.", + "Aliases": [ + "src" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "The name of the new column.", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "TrainingWindowSize", + "Type": "Int", + "Desc": "The number of points from the beginning of the sequence used for training.", + "Aliases": [ + "twnd" + ], + "Required": true, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Confidence", + "Type": "Float", + "Desc": "The confidence for spike detection in the range [0, 100].", + "Aliases": [ + "cnf" + ], + "Required": true, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 99.0 + }, + { + "Name": "SeasonalWindowSize", + "Type": "Int", + "Desc": "An upper bound on the largest relevant seasonality in the input time-series.", + "Aliases": [ + "swnd" + ], + "Required": true, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "Side", + "Type": { + "Kind": "Enum", + "Values": [ + "Positive", + "Negative", + "TwoSided" + ] + }, + "Desc": "The argument that determines whether to detect positive or negative anomalies, or both.", + "Aliases": [ + "side" + ], + "Required": false, + "SortOrder": 101.0, + "IsNullable": false, + "Default": "TwoSided" + }, + { + "Name": "PvalueHistoryLength", + "Type": "Int", + "Desc": "The size of the sliding window for computing the p-value.", + "Aliases": [ + "wnd" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "ErrorFunction", + "Type": { + "Kind": "Enum", + "Values": [ + "SignedDifference", + "AbsoluteDifference", + "SignedProportion", + "AbsoluteProportion", + "SquaredDifference" + ] + }, + "Desc": "The function used to compute the error between the expected and the observed value.", + "Aliases": [ + "err" + ], + "Required": false, + "SortOrder": 103.0, + "IsNullable": false, + "Default": "SignedDifference" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Trainers.AveragedPerceptronBinaryClassifier", + "Desc": "Averaged Perceptron Binary Classifier.", + "FriendlyName": "Averaged Perceptron", + "ShortName": "ap", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "ClassificationLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "HingeLoss" + } + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.01, + 0.1, + 0.5, + 1.0 + ] + } + }, + { + "Name": "DecreaseLearningRate", + "Type": "Bool", + "Desc": "Decrease learning rate", + "Aliases": [ + "decreaselr" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 Regularization Weight", + "Aliases": [ + "reg", + "L2RegularizerWeight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 0.4 + } + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter", + "numIterations" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 100, + "StepSize": 10.0, + "IsLogScale": true + } + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "initWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "ResetWeightsAfterXExamples", + "Type": "Int", + "Desc": "Number of examples after which weights will be reset to the current average", + "Aliases": [ + "numreset" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LazyUpdate", + "Type": "Bool", + "Desc": "Instead of updating averaged weights on every example, only update when loss is nonzero", + "Aliases": [ + "lazy", + "DoLazyUpdates" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "RecencyGain", + "Type": "Float", + "Desc": "Extra weight given to more recent updates", + "Aliases": [ + "rg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "RecencyGainMultiplicative", + "Type": "Bool", + "Desc": "Whether Recency Gain is multiplicative (vs. additive)", + "Aliases": [ + "rgm", + "RecencyGainMulti" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Averaged", + "Type": "Bool", + "Desc": "Do averaging?", + "Aliases": [ + "avg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "AveragedTolerance", + "Type": "Float", + "Desc": "The inexactness tolerance for averaging", + "Aliases": [ + "avgtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01 + }, + { + "Name": "InitialWeights", + "Type": "String", + "Desc": "Initial Weights and bias, comma-separated", + "Aliases": [ + "initweights" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Whether to shuffle for each training iteration", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.EnsembleBinaryClassifier", + "Desc": "Train binary ensemble.", + "FriendlyName": "Parallel Ensemble (bagging, stacking, etc)", + "ShortName": null, + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "SamplingType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleSubsetSelector" + }, + "Desc": "Sampling Type", + "Aliases": [ + "st" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": { + "Name": "BootstrapSelector", + "Settings": { + "FeatureSelector": { + "Name": "AllFeatureSelector" + } + } + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "NumModels", + "Type": "Int", + "Desc": "Number of models per batch. If not specified, will default to 50 if there is only one base predictor, or the number of base predictors otherwise.", + "Aliases": [ + "nm" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "SubModelSelectorType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleBinarySubModelSelector" + }, + "Desc": "Algorithm to prune the base learners for selective Ensemble", + "Aliases": [ + "pt" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": { + "Name": "AllSelector" + } + }, + { + "Name": "OutputCombiner", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleBinaryOutputCombiner" + }, + "Desc": "Output combiner", + "Aliases": [ + "oc" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": { + "Name": "Median" + } + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "TrainParallel", + "Type": "Bool", + "Desc": "All the base learners will run asynchronously if the value is true", + "Aliases": [ + "tp" + ], + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Batch size", + "Aliases": [ + "bs" + ], + "Required": false, + "SortOrder": 107.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "ShowMetrics", + "Type": "Bool", + "Desc": "True, if metrics for each model need to be evaluated and shown in comparison table. This is done by using validation set if available or the training set", + "Aliases": [ + "sm" + ], + "Required": false, + "SortOrder": 108.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.EnsembleClassification", + "Desc": "Train multiclass ensemble.", + "FriendlyName": "Parallel Ensemble (bagging, stacking, etc)", + "ShortName": null, + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "SamplingType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleSubsetSelector" + }, + "Desc": "Sampling Type", + "Aliases": [ + "st" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": { + "Name": "BootstrapSelector", + "Settings": { + "FeatureSelector": { + "Name": "AllFeatureSelector" + } + } + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "NumModels", + "Type": "Int", + "Desc": "Number of models per batch. If not specified, will default to 50 if there is only one base predictor, or the number of base predictors otherwise.", + "Aliases": [ + "nm" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "SubModelSelectorType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleMulticlassSubModelSelector" + }, + "Desc": "Algorithm to prune the base learners for selective Ensemble", + "Aliases": [ + "pt" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": { + "Name": "AllSelectorMultiClass" + } + }, + { + "Name": "OutputCombiner", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleMulticlassOutputCombiner" + }, + "Desc": "Output combiner", + "Aliases": [ + "oc" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": { + "Name": "MultiMedian" + } + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "TrainParallel", + "Type": "Bool", + "Desc": "All the base learners will run asynchronously if the value is true", + "Aliases": [ + "tp" + ], + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Batch size", + "Aliases": [ + "bs" + ], + "Required": false, + "SortOrder": 107.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "ShowMetrics", + "Type": "Bool", + "Desc": "True, if metrics for each model need to be evaluated and shown in comparison table. This is done by using validation set if available or the training set", + "Aliases": [ + "sm" + ], + "Required": false, + "SortOrder": 108.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.EnsembleRegression", + "Desc": "Train regression ensemble.", + "FriendlyName": "Parallel Ensemble (bagging, stacking, etc)", + "ShortName": null, + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "SamplingType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleSubsetSelector" + }, + "Desc": "Sampling Type", + "Aliases": [ + "st" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": { + "Name": "BootstrapSelector", + "Settings": { + "FeatureSelector": { + "Name": "AllFeatureSelector" + } + } + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "NumModels", + "Type": "Int", + "Desc": "Number of models per batch. If not specified, will default to 50 if there is only one base predictor, or the number of base predictors otherwise.", + "Aliases": [ + "nm" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "SubModelSelectorType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleRegressionSubModelSelector" + }, + "Desc": "Algorithm to prune the base learners for selective Ensemble", + "Aliases": [ + "pt" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": { + "Name": "AllSelector" + } + }, + { + "Name": "OutputCombiner", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleRegressionOutputCombiner" + }, + "Desc": "Output combiner", + "Aliases": [ + "oc" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": { + "Name": "Median" + } + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "TrainParallel", + "Type": "Bool", + "Desc": "All the base learners will run asynchronously if the value is true", + "Aliases": [ + "tp" + ], + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Batch size", + "Aliases": [ + "bs" + ], + "Required": false, + "SortOrder": 107.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "ShowMetrics", + "Type": "Bool", + "Desc": "True, if metrics for each model need to be evaluated and shown in comparison table. This is done by using validation set if available or the training set", + "Aliases": [ + "sm" + ], + "Required": false, + "SortOrder": 108.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastForestBinaryClassifier", + "Desc": "Uses a random forest learner to perform binary classification.", + "FriendlyName": "Fast Forest Classification", + "ShortName": "ff", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "MaximumOutputMagnitudePerTree", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "NumberOfQuantileSamples", + "Type": "Int", + "Desc": "Number of labels to be sampled from each leaf to make the distribution", + "Aliases": [ + "qsc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastForestRegressor", + "Desc": "Trains a random forest to fit target values using least-squares.", + "FriendlyName": "FastForestRegression", + "ShortName": "ffr", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "ShuffleLabels", + "Type": "Bool", + "Desc": "Shuffle the labels on every iteration. Useful probably only if using this tree as a tree leaf featurizer for multiclass.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NumberOfQuantileSamples", + "Type": "Int", + "Desc": "Number of labels to be sampled from each leaf to make the distribution", + "Aliases": [ + "qsc" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastTreeBinaryClassifier", + "Desc": "Uses a logit-boost boosted tree learner to perform binary classification.", + "FriendlyName": "FastTree (Boosted Trees) Classification", + "ShortName": "ftc", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "UnbalancedSets", + "Type": "Bool", + "Desc": "Option for using derivatives optimized for unbalanced sets", + "Aliases": [ + "us" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastTreeRanker", + "Desc": "Trains gradient boosted decision trees to the LambdaRank quasi-gradient.", + "FriendlyName": "FastTree (Boosted Trees) Ranking", + "ShortName": "ftrank", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "CustomGains", + "Type": { + "Kind": "Array", + "ItemType": "Float" + }, + "Desc": "Comma-separated list of gains associated to each relevance label.", + "Aliases": [ + "gains" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + 0.0, + 3.0, + 7.0, + 15.0, + 31.0 + ] + }, + { + "Name": "UseDcg", + "Type": "Bool", + "Desc": "Train DCG instead of NDCG", + "Aliases": [ + "dcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "SortingAlgorithm", + "Type": "String", + "Desc": "The sorting algorithm to use for DCG and LambdaMart calculations [DescendingStablePessimistic/DescendingStable/DescendingReverse/DescendingDotNet]", + "Aliases": [ + "sort" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "DescendingStablePessimistic" + }, + { + "Name": "NdcgTruncationLevel", + "Type": "Int", + "Desc": "max-NDCG truncation to use in the LambdaMART algorithm", + "Aliases": [ + "n" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "ShiftedNdcg", + "Type": "Bool", + "Desc": "Use shifted NDCG", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "CostFunctionParam", + "Type": "Char", + "Desc": "Cost function parameter (w/c)", + "Aliases": [ + "cf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "w" + }, + { + "Name": "DistanceWeight2", + "Type": "Bool", + "Desc": "Distance weight 2 adjustment to cost", + "Aliases": [ + "dw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NormalizeQueryLambdas", + "Type": "Bool", + "Desc": "Normalize query lambdas", + "Aliases": [ + "nql" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRankingOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastTreeRegressor", + "Desc": "Trains gradient boosted decision trees to fit target values using least-squares.", + "FriendlyName": "FastTree (Boosted Trees) Regression", + "ShortName": "ftr", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FastTreeTweedieRegressor", + "Desc": "Trains gradient boosted decision trees to fit target values using a Tweedie loss function. This learner is a generalization of Poisson, compound Poisson, and gamma regression.", + "FriendlyName": "FastTree (Boosted Trees) Tweedie Regression", + "ShortName": "fttweedie", + "Inputs": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Index", + "Type": "Float", + "Desc": "Index parameter for the Tweedie distribution, in the range [1, 2]. 1 is Poisson loss, 2 is gamma loss, and intermediate values are compound Poisson loss.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.5 + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.FieldAwareFactorizationMachineBinaryClassifier", + "Desc": "Train a field-aware factorization machine for binary classification", + "FriendlyName": "Field-aware Factorization Machine", + "ShortName": "ffm", + "Inputs": [ + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Initial learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 0.1, + "SweepRange": { + "RangeType": "Float", + "Min": 0.001, + "Max": 1.0, + "IsLogScale": true + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of training iterations", + "Aliases": [ + "iters", + "iter" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 5, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 100 + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LatentDimension", + "Type": "Int", + "Desc": "Latent space dimension", + "Aliases": [ + "d" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 4, + "Max": 100 + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LambdaLinear", + "Type": "Float", + "Desc": "Regularization coefficient of linear weights", + "Aliases": [ + "lambdaLinear" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.0001, + "SweepRange": { + "RangeType": "Float", + "Min": 1E-08, + "Max": 1.0, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "LambdaLatent", + "Type": "Float", + "Desc": "Regularization coefficient of latent weights", + "Aliases": [ + "lambdaLatent" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 0.0001, + "SweepRange": { + "RangeType": "Float", + "Min": 1E-08, + "Max": 1.0, + "IsLogScale": true + } + }, + { + "Name": "NormalizeFeatures", + "Type": "Bool", + "Desc": "Whether to normalize the input vectors so that the concatenation of all fields' feature vectors is unit-length", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "ExtraFeatureColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Extra columns to use for feature vectors. The i-th specified string denotes the column containing features form the (i+1)-th field. Note that the first field is specified by \"feat\" instead of \"exfeat\".", + "Aliases": [ + "exfeat" + ], + "Required": false, + "SortOrder": 7.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Whether to shuffle for each training iteration", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 90.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Verbose", + "Type": "Bool", + "Desc": "Report traning progress or not", + "Aliases": [ + "verbose" + ], + "Required": false, + "SortOrder": 91.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Radius", + "Type": "Float", + "Desc": "Radius of initial latent factors", + "Aliases": [ + "rad" + ], + "Required": false, + "SortOrder": 110.0, + "IsNullable": false, + "Default": 0.5, + "SweepRange": { + "RangeType": "Float", + "Min": 0.1, + "Max": 1.0 + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.GeneralizedAdditiveModelBinaryClassifier", + "Desc": "Trains a gradient boosted stump per feature, on all features simultaneously, to fit target values using least-squares. It maintains no interactions between features.", + "FriendlyName": "Generalized Additive Model for Binary Classification", + "ShortName": "gam", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Total number of iterations over all features", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 9500, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 200, + 1500, + 9500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of training instances required to form a partition", + "Aliases": [ + "mi" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.002, + "SweepRange": { + "RangeType": "Float", + "Min": 0.001, + "Max": 0.1, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "UnbalancedSets", + "Type": "Bool", + "Desc": "Should we use derivatives optimized for unbalanced sets", + "Aliases": [ + "us" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Int", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Infinity" + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.GeneralizedAdditiveModelRegressor", + "Desc": "Trains a gradient boosted stump per feature, on all features simultaneously, to fit target values using least-squares. It maintains no interactions between features.", + "FriendlyName": "Generalized Additive Model for Regression", + "ShortName": "gamr", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Total number of iterations over all features", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 9500, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 200, + 1500, + 9500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of training instances required to form a partition", + "Aliases": [ + "mi" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.002, + "SweepRange": { + "RangeType": "Float", + "Min": 0.001, + "Max": 0.1, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "PruningMetrics", + "Type": "Int", + "Desc": "Metric for pruning. (For regression, 1: L1, 2:L2; default L2)", + "Aliases": [ + "pmetric" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Int", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Infinity" + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.KMeansPlusPlusClusterer", + "Desc": "K-means is a popular clustering algorithm. With K-means, the data is clustered into a specified number of clusters in order to minimize the within-cluster sum of squares. K-means++ improves upon K-means by using a better method for choosing the initial cluster centers.", + "FriendlyName": "KMeans++ Clustering", + "ShortName": "KM", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "K", + "Type": "Int", + "Desc": "The number of clusters", + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 5, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 5, + 10, + 20, + 40 + ] + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Defaults to automatic. Determinism not guaranteed.", + "Aliases": [ + "nt", + "t", + "threads" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "InitializationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "KMeansPlusPlus", + "Random", + "KMeansYinyang" + ] + }, + "Desc": "Cluster initialization algorithm", + "Aliases": [ + "init" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "KMeansYinyang" + }, + { + "Name": "OptTol", + "Type": "Float", + "Desc": "Tolerance parameter for trainer convergence. Low = slower, more accurate", + "Aliases": [ + "ot" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1E-07 + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum number of iterations.", + "Aliases": [ + "maxiter", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000 + }, + { + "Name": "AccelMemBudgetMb", + "Type": "Int", + "Desc": "Memory budget (in MBs) to use for KMeans acceleration", + "Aliases": [ + "accelMemBudgetMb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 4096 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "IUnsupervisedTrainerWithWeight", + "ITrainerInput" + ], + "OutputKind": [ + "IClusteringOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LightGbmBinaryClassifier", + "Desc": "Train a LightGBM binary classification model.", + "FriendlyName": "LightGBM Binary Classifier", + "ShortName": "LightGBM", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations.", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 50, + 100, + 150, + 200 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Shrinkage rate for trees, used to prevent over-fitting. Range: (0,1].", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "Maximum leaves for trees.", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of instances needed in a child.", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20, + 50 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "Booster", + "Type": { + "Kind": "Component", + "ComponentKind": "BoosterParameterFunction" + }, + "Desc": "Which booster to use, can be gbtree, gblinear or dart. gbtree and dart use tree based model while gblinear uses linear function.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": { + "Name": "gbdt" + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "UnbalancedSets", + "Type": "Bool", + "Desc": "Use for binary classification when training data is not balanced.", + "Aliases": [ + "us" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "WeightOfPositiveExamples", + "Type": "Float", + "Desc": "Control the balance of positive and negative weights, useful for unbalanced classes. A typical value to consider: sum(negative cases) / sum(positive cases).", + "Aliases": [ + "ScalePosWeight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Sigmoid", + "Type": "Float", + "Desc": "Parameter for the sigmoid function.", + "Aliases": [ + "sigmoid" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "EvaluationMetric", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Default", + "Logloss", + "Error", + "AreaUnderCurve" + ] + }, + "Desc": "Evaluation metrics.", + "Aliases": [ + "em" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Logloss" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of bucket bin for features.", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "Verbose", + "Type": "Bool", + "Desc": "Verbose", + "Aliases": [ + "v" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Silent", + "Type": "Bool", + "Desc": "Printing running messages.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of parallel threads used to run LightGBM.", + "Aliases": [ + "nt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "EarlyStoppingRound", + "Type": "Int", + "Desc": "Rounds of early stopping, 0 will disable it.", + "Aliases": [ + "es" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of entries in a batch when loading data.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1048576 + }, + { + "Name": "UseCategoricalSplit", + "Type": "Bool", + "Desc": "Enable categorical split or not.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "HandleMissingValue", + "Type": "Bool", + "Desc": "Enable special handling of missing value or not.", + "Aliases": [ + "hmv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "UseZeroAsMissingValue", + "Type": "Bool", + "Desc": "Enable usage of zero (0) as missing value.", + "Aliases": [ + "uzam" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "MinimumExampleCountPerGroup", + "Type": "Int", + "Desc": "Minimum number of instances per categorical group.", + "Aliases": [ + "mdpg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 50, + 100, + 200 + ] + } + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Max number of categorical thresholds.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 32, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 8, + 16, + 32, + 64 + ] + } + }, + { + "Name": "CategoricalSmoothing", + "Type": "Float", + "Desc": "Lapalace smooth term in categorical feature spilt. Avoid the bias of small categories.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20 + ] + } + }, + { + "Name": "L2CategoricalRegularization", + "Type": "Float", + "Desc": "L2 Regularization for categorical split.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.5, + 1, + 5, + 10 + ] + } + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Sets the random seed for LightGBM to use.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelLightGBM" + }, + "Desc": "Parallel LightGBM Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LightGbmClassifier", + "Desc": "Train a LightGBM multi class model.", + "FriendlyName": "LightGBM Multi Class Classifier", + "ShortName": "LightGBMMC", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations.", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 50, + 100, + 150, + 200 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Shrinkage rate for trees, used to prevent over-fitting. Range: (0,1].", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "Maximum leaves for trees.", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of instances needed in a child.", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20, + 50 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "Booster", + "Type": { + "Kind": "Component", + "ComponentKind": "BoosterParameterFunction" + }, + "Desc": "Which booster to use, can be gbtree, gblinear or dart. gbtree and dart use tree based model while gblinear uses linear function.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": { + "Name": "gbdt" + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "UnbalancedSets", + "Type": "Bool", + "Desc": "Use for multi-class classification when training data is not balanced", + "Aliases": [ + "us" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseSoftmax", + "Type": "Bool", + "Desc": "Use softmax loss for the multi classification.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "Sigmoid", + "Type": "Float", + "Desc": "Parameter for the sigmoid function.", + "Aliases": [ + "sigmoid" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "EvaluationMetric", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Default", + "Error", + "LogLoss" + ] + }, + "Desc": "Evaluation metrics.", + "Aliases": [ + "em" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Error" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of bucket bin for features.", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "Verbose", + "Type": "Bool", + "Desc": "Verbose", + "Aliases": [ + "v" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Silent", + "Type": "Bool", + "Desc": "Printing running messages.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of parallel threads used to run LightGBM.", + "Aliases": [ + "nt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "EarlyStoppingRound", + "Type": "Int", + "Desc": "Rounds of early stopping, 0 will disable it.", + "Aliases": [ + "es" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of entries in a batch when loading data.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1048576 + }, + { + "Name": "UseCategoricalSplit", + "Type": "Bool", + "Desc": "Enable categorical split or not.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "HandleMissingValue", + "Type": "Bool", + "Desc": "Enable special handling of missing value or not.", + "Aliases": [ + "hmv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "UseZeroAsMissingValue", + "Type": "Bool", + "Desc": "Enable usage of zero (0) as missing value.", + "Aliases": [ + "uzam" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "MinimumExampleCountPerGroup", + "Type": "Int", + "Desc": "Minimum number of instances per categorical group.", + "Aliases": [ + "mdpg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 50, + 100, + 200 + ] + } + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Max number of categorical thresholds.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 32, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 8, + 16, + 32, + 64 + ] + } + }, + { + "Name": "CategoricalSmoothing", + "Type": "Float", + "Desc": "Lapalace smooth term in categorical feature spilt. Avoid the bias of small categories.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20 + ] + } + }, + { + "Name": "L2CategoricalRegularization", + "Type": "Float", + "Desc": "L2 Regularization for categorical split.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.5, + 1, + 5, + 10 + ] + } + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Sets the random seed for LightGBM to use.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelLightGBM" + }, + "Desc": "Parallel LightGBM Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LightGbmRanker", + "Desc": "Train a LightGBM ranking model.", + "FriendlyName": "LightGBM Ranking", + "ShortName": "LightGBMRank", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations.", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 50, + 100, + 150, + 200 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Shrinkage rate for trees, used to prevent over-fitting. Range: (0,1].", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "Maximum leaves for trees.", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of instances needed in a child.", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20, + 50 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "Booster", + "Type": { + "Kind": "Component", + "ComponentKind": "BoosterParameterFunction" + }, + "Desc": "Which booster to use, can be gbtree, gblinear or dart. gbtree and dart use tree based model while gblinear uses linear function.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": { + "Name": "gbdt" + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "CustomGains", + "Type": { + "Kind": "Array", + "ItemType": "Int" + }, + "Desc": "An array of gains associated to each relevance label.", + "Aliases": [ + "gains" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + 0, + 3, + 7, + 15, + 31, + 63, + 127, + 255, + 511, + 1023, + 2047, + 4095 + ] + }, + { + "Name": "Sigmoid", + "Type": "Float", + "Desc": "Parameter for the sigmoid function.", + "Aliases": [ + "sigmoid" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "EvaluationMetric", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Default", + "MeanAveragedPrecision", + "NormalizedDiscountedCumulativeGain" + ] + }, + "Desc": "Evaluation metrics.", + "Aliases": [ + "em" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "NormalizedDiscountedCumulativeGain" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of bucket bin for features.", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "Verbose", + "Type": "Bool", + "Desc": "Verbose", + "Aliases": [ + "v" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Silent", + "Type": "Bool", + "Desc": "Printing running messages.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of parallel threads used to run LightGBM.", + "Aliases": [ + "nt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "EarlyStoppingRound", + "Type": "Int", + "Desc": "Rounds of early stopping, 0 will disable it.", + "Aliases": [ + "es" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of entries in a batch when loading data.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1048576 + }, + { + "Name": "UseCategoricalSplit", + "Type": "Bool", + "Desc": "Enable categorical split or not.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "HandleMissingValue", + "Type": "Bool", + "Desc": "Enable special handling of missing value or not.", + "Aliases": [ + "hmv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "UseZeroAsMissingValue", + "Type": "Bool", + "Desc": "Enable usage of zero (0) as missing value.", + "Aliases": [ + "uzam" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "MinimumExampleCountPerGroup", + "Type": "Int", + "Desc": "Minimum number of instances per categorical group.", + "Aliases": [ + "mdpg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 50, + 100, + 200 + ] + } + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Max number of categorical thresholds.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 32, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 8, + 16, + 32, + 64 + ] + } + }, + { + "Name": "CategoricalSmoothing", + "Type": "Float", + "Desc": "Lapalace smooth term in categorical feature spilt. Avoid the bias of small categories.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20 + ] + } + }, + { + "Name": "L2CategoricalRegularization", + "Type": "Float", + "Desc": "L2 Regularization for categorical split.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.5, + 1, + 5, + 10 + ] + } + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Sets the random seed for LightGBM to use.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelLightGBM" + }, + "Desc": "Parallel LightGBM Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRankingOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LightGbmRegressor", + "Desc": "LightGBM Regression", + "FriendlyName": "LightGBM Regressor", + "ShortName": "LightGBMR", + "Inputs": [ + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations.", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 50, + 100, + 150, + 200 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Shrinkage rate for trees, used to prevent over-fitting. Range: (0,1].", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "Maximum leaves for trees.", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "Minimum number of instances needed in a child.", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20, + 50 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "Booster", + "Type": { + "Kind": "Component", + "ComponentKind": "BoosterParameterFunction" + }, + "Desc": "Which booster to use, can be gbtree, gblinear or dart. gbtree and dart use tree based model while gblinear uses linear function.", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": { + "Name": "gbdt" + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "EvaluationMetric", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "Default", + "MeanAbsoluteError", + "RootMeanSquaredError", + "MeanSquaredError" + ] + }, + "Desc": "Evaluation metrics.", + "Aliases": [ + "em" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "RootMeanSquaredError" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of bucket bin for features.", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "Verbose", + "Type": "Bool", + "Desc": "Verbose", + "Aliases": [ + "v" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Silent", + "Type": "Bool", + "Desc": "Printing running messages.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of parallel threads used to run LightGBM.", + "Aliases": [ + "nt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "EarlyStoppingRound", + "Type": "Int", + "Desc": "Rounds of early stopping, 0 will disable it.", + "Aliases": [ + "es" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of entries in a batch when loading data.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1048576 + }, + { + "Name": "UseCategoricalSplit", + "Type": "Bool", + "Desc": "Enable categorical split or not.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "HandleMissingValue", + "Type": "Bool", + "Desc": "Enable special handling of missing value or not.", + "Aliases": [ + "hmv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "UseZeroAsMissingValue", + "Type": "Bool", + "Desc": "Enable usage of zero (0) as missing value.", + "Aliases": [ + "uzam" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + true, + false + ] + } + }, + { + "Name": "MinimumExampleCountPerGroup", + "Type": "Int", + "Desc": "Minimum number of instances per categorical group.", + "Aliases": [ + "mdpg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 50, + 100, + 200 + ] + } + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Max number of categorical thresholds.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 32, + "Range": { + "Inf": 0, + "Max": 2147483647 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 8, + 16, + 32, + 64 + ] + } + }, + { + "Name": "CategoricalSmoothing", + "Type": "Float", + "Desc": "Lapalace smooth term in categorical feature spilt. Avoid the bias of small categories.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 20 + ] + } + }, + { + "Name": "L2CategoricalRegularization", + "Type": "Float", + "Desc": "L2 Regularization for categorical split.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.5, + 1, + 5, + 10 + ] + } + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Sets the random seed for LightGBM to use.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelLightGBM" + }, + "Desc": "Parallel LightGBM Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithGroupId", + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LinearSvmBinaryClassifier", + "Desc": "Train a linear SVM.", + "FriendlyName": "SVM (Pegasos-Linear)", + "ShortName": "svm", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight", + "WeightColumn" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Lambda", + "Type": "Float", + "Desc": "Regularizer constant", + "Aliases": [ + "lambda" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.001, + "SweepRange": { + "RangeType": "Float", + "Min": 1E-05, + "Max": 0.1, + "StepSize": 10.0, + "IsLogScale": true + } + }, + { + "Name": "PerformProjection", + "Type": "Bool", + "Desc": "Perform projection to unit-ball? Typically used with batch size > 1.", + "Aliases": [ + "project" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter", + "numIterations" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 100, + "StepSize": 10.0, + "IsLogScale": true + } + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "initWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "NoBias", + "Type": "Bool", + "Desc": "No bias", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "InitialWeights", + "Type": "String", + "Desc": "Initial Weights and bias, comma-separated", + "Aliases": [ + "initweights" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Whether to shuffle for each training iteration", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Batch size", + "Aliases": [ + "batch" + ], + "Required": false, + "SortOrder": 190.0, + "IsNullable": false, + "Default": 1 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LocalDeepSvmBinaryClassifier", + "Desc": "LD-SVM learns a binary, non-linear SVM classifier with a kernel that is specifically designed to reduce prediction time. LD-SVM learns decision boundaries that are locally linear.", + "FriendlyName": "Local Deep SVM (LDSVM)", + "ShortName": "LDSVM", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "TreeDepth", + "Type": "Int", + "Desc": "Depth of Local Deep SVM tree", + "Aliases": [ + "depth" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 3, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 3, + 5, + 7 + ] + } + }, + { + "Name": "LambdaW", + "Type": "Float", + "Desc": "Regularizer for classifier parameter W", + "Aliases": [ + "lw" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.1, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.01, + 0.001 + ] + } + }, + { + "Name": "LambdaTheta", + "Type": "Float", + "Desc": "Regularizer for kernel parameter Theta", + "Aliases": [ + "lt" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.01, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.01, + 0.001 + ] + } + }, + { + "Name": "LambdaThetaprime", + "Type": "Float", + "Desc": "Regularizer for kernel parameter Thetaprime", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.01, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.1, + 0.01, + 0.001 + ] + } + }, + { + "Name": "Sigma", + "Type": "Float", + "Desc": "Parameter for sigmoid sharpness", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1.0, + 0.1, + 0.01 + ] + } + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter", + "NumIterations" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 15000, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10000, + 15000 + ] + } + }, + { + "Name": "UseBias", + "Type": "Bool", + "Desc": "No bias", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "Cache", + "Type": "Bool", + "Desc": "Whether to cache the data before the first iteration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LogisticRegressionBinaryClassifier", + "Desc": "Logistic Regression is a method in statistics used to predict the probability of occurrence of an event and can be used as a classification algorithm. The algorithm predicts the probability of occurrence of an event by fitting data to a logistical function.", + "FriendlyName": "Logistic Regression", + "ShortName": "lr", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "ShowTrainingStatistics", + "Type": "Bool", + "Desc": "Show statistics of training examples.", + "Aliases": [ + "stat", + "ShowTrainingStats" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization weight", + "Aliases": [ + "l2", + "L2Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization weight", + "Aliases": [ + "l1", + "L1Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "OptimizationTolerance", + "Type": "Float", + "Desc": "Tolerance parameter for optimization convergence. Low = slower, more accurate", + "Aliases": [ + "ot", + "OptTol" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1E-07, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0001, + 1E-07 + ] + } + }, + { + "Name": "HistorySize", + "Type": "Int", + "Desc": "Memory size for L-BFGS. Low=faster, less accurate", + "Aliases": [ + "m", + "MemorySize" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 5, + 20, + 50 + ] + } + }, + { + "Name": "EnforceNonNegativity", + "Type": "Bool", + "Desc": "Enforce non-negative weights", + "Aliases": [ + "nn" + ], + "Required": false, + "SortOrder": 90.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "InitWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum iterations.", + "Aliases": [ + "maxiter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 2147483647 + } + }, + { + "Name": "StochasticGradientDescentInitilaizationTolerance", + "Type": "Float", + "Desc": "Run SGD to initialize LR weights, converging to this tolerance", + "Aliases": [ + "sgd", + "SgdInitializationTolerance" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Quiet", + "Type": "Bool", + "Desc": "If set to true, produce no output during training.", + "Aliases": [ + "q" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseThreads", + "Type": "Bool", + "Desc": "Whether or not to use threads. Default is true", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of threads", + "Aliases": [ + "nt", + "NumThreads" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DenseOptimizer", + "Type": "Bool", + "Desc": "Force densification of the internal optimization vectors", + "Aliases": [ + "do" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.LogisticRegressionClassifier", + "Desc": "Maximum entropy classification is a method in statistics used to predict the probabilities of parallel events. The model predicts the probabilities of parallel events by fitting data to a softmax function.", + "FriendlyName": "Multi-class Logistic Regression", + "ShortName": "mlr", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "ShowTrainingStatistics", + "Type": "Bool", + "Desc": "Show statistics of training examples.", + "Aliases": [ + "stat", + "ShowTrainingStats" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization weight", + "Aliases": [ + "l2", + "L2Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization weight", + "Aliases": [ + "l1", + "L1Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "OptimizationTolerance", + "Type": "Float", + "Desc": "Tolerance parameter for optimization convergence. Low = slower, more accurate", + "Aliases": [ + "ot", + "OptTol" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1E-07, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0001, + 1E-07 + ] + } + }, + { + "Name": "HistorySize", + "Type": "Int", + "Desc": "Memory size for L-BFGS. Low=faster, less accurate", + "Aliases": [ + "m", + "MemorySize" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 5, + 20, + 50 + ] + } + }, + { + "Name": "EnforceNonNegativity", + "Type": "Bool", + "Desc": "Enforce non-negative weights", + "Aliases": [ + "nn" + ], + "Required": false, + "SortOrder": 90.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "InitWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum iterations.", + "Aliases": [ + "maxiter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 2147483647 + } + }, + { + "Name": "StochasticGradientDescentInitilaizationTolerance", + "Type": "Float", + "Desc": "Run SGD to initialize LR weights, converging to this tolerance", + "Aliases": [ + "sgd", + "SgdInitializationTolerance" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Quiet", + "Type": "Bool", + "Desc": "If set to true, produce no output during training.", + "Aliases": [ + "q" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseThreads", + "Type": "Bool", + "Desc": "Whether or not to use threads. Default is true", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of threads", + "Aliases": [ + "nt", + "NumThreads" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DenseOptimizer", + "Type": "Bool", + "Desc": "Force densification of the internal optimization vectors", + "Aliases": [ + "do" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.NaiveBayesClassifier", + "Desc": "Train a MulticlassNaiveBayesTrainer.", + "FriendlyName": "Multiclass Naive Bayes", + "ShortName": "MNB", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.OnlineGradientDescentRegressor", + "Desc": "Train a Online gradient descent perceptron.", + "FriendlyName": "Stochastic Gradient Descent (Regression)", + "ShortName": "ogd", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "RegressionLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "SquaredLoss" + } + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.1, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.01, + 0.1, + 0.5, + 1.0 + ] + } + }, + { + "Name": "DecreaseLearningRate", + "Type": "Bool", + "Desc": "Decrease learning rate", + "Aliases": [ + "decreaselr" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 Regularization Weight", + "Aliases": [ + "reg", + "L2RegularizerWeight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 0.4 + } + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter", + "numIterations" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 100, + "StepSize": 10.0, + "IsLogScale": true + } + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "initWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "ResetWeightsAfterXExamples", + "Type": "Int", + "Desc": "Number of examples after which weights will be reset to the current average", + "Aliases": [ + "numreset" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LazyUpdate", + "Type": "Bool", + "Desc": "Instead of updating averaged weights on every example, only update when loss is nonzero", + "Aliases": [ + "lazy", + "DoLazyUpdates" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "RecencyGain", + "Type": "Float", + "Desc": "Extra weight given to more recent updates", + "Aliases": [ + "rg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "RecencyGainMultiplicative", + "Type": "Bool", + "Desc": "Whether Recency Gain is multiplicative (vs. additive)", + "Aliases": [ + "rgm", + "RecencyGainMulti" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Averaged", + "Type": "Bool", + "Desc": "Do averaging?", + "Aliases": [ + "avg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "AveragedTolerance", + "Type": "Float", + "Desc": "The inexactness tolerance for averaging", + "Aliases": [ + "avgtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01 + }, + { + "Name": "InitialWeights", + "Type": "String", + "Desc": "Initial Weights and bias, comma-separated", + "Aliases": [ + "initweights" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Whether to shuffle for each training iteration", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.OrdinaryLeastSquaresRegressor", + "Desc": "Train an OLS regression model.", + "FriendlyName": "Ordinary Least Squares (Regression)", + "ShortName": "ols", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization weight", + "Aliases": [ + "l2" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1E-06, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1E-06, + 0.1, + 1.0 + ] + } + }, + { + "Name": "CalculateStatistics", + "Type": "Bool", + "Desc": "Whether to calculate per parameter significance statistics", + "Aliases": [ + "sig" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of entries in a batch when loading data (0 = auto).", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.PcaAnomalyDetector", + "Desc": "Train an PCA Anomaly model.", + "FriendlyName": "PCA Anomaly Detector", + "ShortName": "pcaAnom", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Rank", + "Type": "Int", + "Desc": "The number of components in the PCA", + "Aliases": [ + "k" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 40, + 80 + ] + } + }, + { + "Name": "Oversampling", + "Type": "Int", + "Desc": "Oversampling parameter for randomized PCA training", + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 40 + ] + } + }, + { + "Name": "Center", + "Type": "Bool", + "Desc": "If enabled, data is centered to be zero mean", + "Aliases": [ + "center" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed for random number generation", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "IUnsupervisedTrainerWithWeight", + "ITrainerInput" + ], + "OutputKind": [ + "IAnomalyDetectionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.PoissonRegressor", + "Desc": "Train an Poisson regression model.", + "FriendlyName": "Poisson Regression", + "ShortName": "PR", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization weight", + "Aliases": [ + "l2", + "L2Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization weight", + "Aliases": [ + "l1", + "L1Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 4 + } + }, + { + "Name": "OptimizationTolerance", + "Type": "Float", + "Desc": "Tolerance parameter for optimization convergence. Low = slower, more accurate", + "Aliases": [ + "ot", + "OptTol" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1E-07, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0001, + 1E-07 + ] + } + }, + { + "Name": "HistorySize", + "Type": "Int", + "Desc": "Memory size for L-BFGS. Low=faster, less accurate", + "Aliases": [ + "m", + "MemorySize" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 5, + 20, + 50 + ] + } + }, + { + "Name": "EnforceNonNegativity", + "Type": "Bool", + "Desc": "Enforce non-negative weights", + "Aliases": [ + "nn" + ], + "Required": false, + "SortOrder": 90.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "InitialWeightsDiameter", + "Type": "Float", + "Desc": "Init weights diameter", + "Aliases": [ + "initwts", + "InitWtsDiameter" + ], + "Required": false, + "SortOrder": 140.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.0, + "Max": 1.0, + "NumSteps": 5 + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum iterations.", + "Aliases": [ + "maxiter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647, + "SweepRange": { + "RangeType": "Long", + "Min": 1, + "Max": 2147483647 + } + }, + { + "Name": "StochasticGradientDescentInitilaizationTolerance", + "Type": "Float", + "Desc": "Run SGD to initialize LR weights, converging to this tolerance", + "Aliases": [ + "sgd", + "SgdInitializationTolerance" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Quiet", + "Type": "Bool", + "Desc": "If set to true, produce no output during training.", + "Aliases": [ + "q" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseThreads", + "Type": "Bool", + "Desc": "Whether or not to use threads. Default is true", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Number of threads", + "Aliases": [ + "nt", + "NumThreads" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DenseOptimizer", + "Type": "Bool", + "Desc": "Force densification of the internal optimization vectors", + "Aliases": [ + "do" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.StochasticDualCoordinateAscentBinaryClassifier", + "Desc": "Train an SDCA binary model.", + "FriendlyName": "Fast Linear (SA-SDCA)", + "ShortName": "SDCA", + "Inputs": [ + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularizer constant. By default the l2 constant is automatically inferred based on data set.", + "Aliases": [ + "l2", + "L2Const" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 1E-07, + 1E-06, + 1E-05, + 0.0001, + 0.001, + 0.01 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "L1Threshold", + "Type": "Float", + "Desc": "L1 soft threshold (L1/L2). Note that it is easier to control and sweep using the threshold parameter than the raw L1-regularizer constant. By default the l1 threshold is automatically inferred based on data set.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 0.0, + 0.25, + 0.5, + 0.75, + 1.0 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "SDCAClassificationLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "LogLoss" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Defaults to automatic. Determinism not guaranteed.", + "Aliases": [ + "nt", + "t", + "threads", + "NumThreads" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "PositiveInstanceWeight", + "Type": "Float", + "Desc": "Apply weight to the positive class, for imbalanced data", + "Aliases": [ + "piw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "ConvergenceTolerance", + "Type": "Float", + "Desc": "The tolerance for the ratio between duality gap and primal loss for convergence checking.", + "Aliases": [ + "tol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.001, + 0.01, + 0.1, + 0.2 + ] + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum number of iterations; set to 1 to simulate online learning. Defaults to automatic.", + "Aliases": [ + "iter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 10, + 20, + 100 + ] + } + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Shuffle data every epoch?", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "ConvergenceCheckFrequency", + "Type": "Int", + "Desc": "Convergence check frequency (in terms of number of iterations). Set as negative or zero for not checking at all. If left blank, it defaults to check after every 'numThreads' iterations.", + "Aliases": [ + "checkFreq", + "CheckFrequency" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "BiasLearningRate", + "Type": "Float", + "Desc": "The learning rate for adjusting bias from being regularized.", + "Aliases": [ + "blr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.01, + 0.1, + 1.0 + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.StochasticDualCoordinateAscentClassifier", + "Desc": "The SDCA linear multi-class classification trainer.", + "FriendlyName": "Fast Linear Multi-class Classification (SA-SDCA)", + "ShortName": "sasdcamc", + "Inputs": [ + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularizer constant. By default the l2 constant is automatically inferred based on data set.", + "Aliases": [ + "l2", + "L2Const" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 1E-07, + 1E-06, + 1E-05, + 0.0001, + 0.001, + 0.01 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "L1Threshold", + "Type": "Float", + "Desc": "L1 soft threshold (L1/L2). Note that it is easier to control and sweep using the threshold parameter than the raw L1-regularizer constant. By default the l1 threshold is automatically inferred based on data set.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 0.0, + 0.25, + 0.5, + 0.75, + 1.0 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "SDCAClassificationLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "LogLoss" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Defaults to automatic. Determinism not guaranteed.", + "Aliases": [ + "nt", + "t", + "threads", + "NumThreads" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ConvergenceTolerance", + "Type": "Float", + "Desc": "The tolerance for the ratio between duality gap and primal loss for convergence checking.", + "Aliases": [ + "tol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.001, + 0.01, + 0.1, + 0.2 + ] + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum number of iterations; set to 1 to simulate online learning. Defaults to automatic.", + "Aliases": [ + "iter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 10, + 20, + 100 + ] + } + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Shuffle data every epoch?", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "ConvergenceCheckFrequency", + "Type": "Int", + "Desc": "Convergence check frequency (in terms of number of iterations). Set as negative or zero for not checking at all. If left blank, it defaults to check after every 'numThreads' iterations.", + "Aliases": [ + "checkFreq", + "CheckFrequency" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "BiasLearningRate", + "Type": "Float", + "Desc": "The learning rate for adjusting bias from being regularized.", + "Aliases": [ + "blr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.01, + 0.1, + 1.0 + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IMulticlassClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.StochasticDualCoordinateAscentRegressor", + "Desc": "The SDCA linear regression trainer.", + "FriendlyName": "Fast Linear Regression (SA-SDCA)", + "ShortName": "sasdcar", + "Inputs": [ + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularizer constant. By default the l2 constant is automatically inferred based on data set.", + "Aliases": [ + "l2", + "L2Const" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 1E-07, + 1E-06, + 1E-05, + 0.0001, + 0.001, + 0.01 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "L1Threshold", + "Type": "Float", + "Desc": "L1 soft threshold (L1/L2). Note that it is easier to control and sweep using the threshold parameter than the raw L1-regularizer constant. By default the l1 threshold is automatically inferred based on data set.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 0.0, + 0.25, + 0.5, + 0.75, + 1.0 + ] + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "SDCARegressionLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "SquaredLoss" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Defaults to automatic. Determinism not guaranteed.", + "Aliases": [ + "nt", + "t", + "threads", + "NumThreads" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ConvergenceTolerance", + "Type": "Float", + "Desc": "The tolerance for the ratio between duality gap and primal loss for convergence checking.", + "Aliases": [ + "tol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.001, + 0.01, + 0.1, + 0.2 + ] + } + }, + { + "Name": "MaximumNumberOfIterations", + "Type": "Int", + "Desc": "Maximum number of iterations; set to 1 to simulate online learning. Defaults to automatic.", + "Aliases": [ + "iter", + "MaxIterations", + "NumberOfIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 10, + 20, + 100 + ] + } + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Shuffle data every epoch?", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "ConvergenceCheckFrequency", + "Type": "Int", + "Desc": "Convergence check frequency (in terms of number of iterations). Set as negative or zero for not checking at all. If left blank, it defaults to check after every 'numThreads' iterations.", + "Aliases": [ + "checkFreq", + "CheckFrequency" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "BiasLearningRate", + "Type": "Float", + "Desc": "The learning rate for adjusting bias from being regularized.", + "Aliases": [ + "blr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.01, + 0.1, + 1.0 + ] + } + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IRegressionOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.StochasticGradientDescentBinaryClassifier", + "Desc": "Train an Hogwild SGD binary model.", + "FriendlyName": "Hogwild SGD (binary)", + "ShortName": "HogwildSGD", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "LossFunction", + "Type": { + "Kind": "Component", + "ComponentKind": "ClassificationLossFunction" + }, + "Desc": "Loss Function", + "Aliases": [ + "loss" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "LogLoss" + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 Regularization constant", + "Aliases": [ + "l2", + "L2Weight" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 1E-06, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1E-07, + 5E-07, + 1E-06, + 5E-06, + 1E-05 + ] + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Defaults to automatic depending on data sparseness. Determinism not guaranteed.", + "Aliases": [ + "nt", + "t", + "threads", + "NumThreads" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Calibrator", + "Type": { + "Kind": "Component", + "ComponentKind": "CalibratorTrainer" + }, + "Desc": "The calibrator kind to apply to the predictor. Specify null for no calibration", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "PlattCalibrator" + } + }, + { + "Name": "MaxCalibrationExamples", + "Type": "Int", + "Desc": "The maximum number of examples to use when training the calibrator", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "ConvergenceTolerance", + "Type": "Float", + "Desc": "Exponential moving averaged improvement tolerance for convergence", + "Aliases": [ + "tol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0001, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.01, + 0.001, + 0.0001, + 1E-05 + ] + } + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Maximum number of iterations; set to 1 to simulate online learning.", + "Aliases": [ + "iter", + "MaxIterations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 5, + 10, + 20 + ] + } + }, + { + "Name": "InitialLearningRate", + "Type": "Float", + "Desc": "Initial learning rate (only used by SGD)", + "Aliases": [ + "ilr", + "lr", + "InitLearningRate" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01 + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Shuffle data every epoch?", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + false, + true + ] + } + }, + { + "Name": "PositiveInstanceWeight", + "Type": "Float", + "Desc": "Apply weight to the positive class, for imbalanced data", + "Aliases": [ + "piw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "CheckFrequency", + "Type": "Int", + "Desc": "Convergence check frequency (in terms of number of iterations). Default equals number of threads", + "Aliases": [ + "checkFreq" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithWeight", + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Trainers.SymSgdBinaryClassifier", + "Desc": "Train a symbolic SGD.", + "FriendlyName": "Symbolic SGD (binary)", + "ShortName": "SymSGD", + "Inputs": [ + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "NumberOfIterations", + "Type": "Int", + "Desc": "Number of passes over the data.", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 50, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 5, + 10, + 20, + 30, + 40, + 50 + ] + } + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "Learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 51.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 10.0, + 1.0, + 0.1, + 0.01, + 0.001 + ] + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization", + "Aliases": [ + "l2" + ], + "Required": false, + "SortOrder": 52.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-05, + 1E-05, + 1E-06, + 1E-07 + ] + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "Degree of lock-free parallelism. Determinism not guaranteed. Multi-threading is not supported currently.", + "Aliases": [ + "nt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Tolerance", + "Type": "Float", + "Desc": "Tolerance for difference in average loss in consecutive passes.", + "Aliases": [ + "tol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0001 + }, + { + "Name": "UpdateFrequency", + "Type": "Int", + "Desc": "The number of iterations each thread learns a local model until combining it with the global model. Low value means more updated global model and high value means less cache traffic.", + "Aliases": [ + "freq" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + "", + 5, + 20 + ] + } + }, + { + "Name": "MemorySize", + "Type": "Int", + "Desc": "The acceleration memory budget in MB", + "Aliases": [ + "accelMemBudget" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1024 + }, + { + "Name": "Shuffle", + "Type": "Bool", + "Desc": "Shuffle data?", + "Aliases": [ + "shuf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "PositiveInstanceWeight", + "Type": "Float", + "Desc": "Apply weight to the positive class, for imbalanced data", + "Aliases": [ + "piw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The trained model" + } + ], + "InputKind": [ + "ITrainerInputWithLabel", + "ITrainerInput" + ], + "OutputKind": [ + "IBinaryClassificationOutput", + "ITrainerOutput" + ] + }, + { + "Name": "Transforms.ApproximateBootstrapSampler", + "Desc": "Approximate bootstrap sampling.", + "FriendlyName": "Bootstrap Sample Transform", + "ShortName": "BootstrapSample", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Complement", + "Type": "Bool", + "Desc": "Whether this is the out-of-bag sample, that is, all those rows that are not selected by the transform.", + "Aliases": [ + "comp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "The random seed. If unspecified random state will be instead derived from the environment.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ShuffleInput", + "Type": "Bool", + "Desc": "Whether we should attempt to shuffle the source data. By default on, but can be turned off for efficiency.", + "Aliases": [ + "si" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "PoolSize", + "Type": "Int", + "Desc": "When shuffling the output, the number of output rows to keep in that pool. Note that shuffling of output is completely distinct from shuffling of input.", + "Aliases": [ + "pool" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.BinaryPredictionScoreColumnsRenamer", + "Desc": "For binary prediction, it renames the PredictedLabel and Score columns to include the name of the positive class.", + "FriendlyName": "Rename Binary Prediction Score Columns", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model used in scoring", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.BinNormalizer", + "Desc": "The values are assigned into equidensity bins and a value is mapped to its bin_number/number_of_bins.", + "FriendlyName": "Binning Normalizer", + "ShortName": "Bin", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "NumBins", + "Type": "Int", + "Desc": "Max number of bins, power of 2 recommended", + "Aliases": [ + "bins" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumBins", + "Type": "Int", + "Desc": "Max number of bins, power of 2 recommended", + "Aliases": [ + "bins" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1024 + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.CategoricalHashOneHotVectorizer", + "Desc": "Converts the categorical value into an indicator array by hashing the value and using the hash as an index in the bag. If the input column is a vector, a single indicator bag is returned for it.", + "FriendlyName": "Categorical Hash Transform", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "OutputKind", + "Type": { + "Kind": "Enum", + "Values": [ + "Bag", + "Indicator", + "Key", + "Binary" + ] + }, + "Desc": "Output kind: Bag (multi-set vector), Ind (indicator vector), or Key (index)", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "The number of bits to hash into. Must be between 1 and 30, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Ordered", + "Type": "Bool", + "Desc": "Whether the position of each term should be included in the hash", + "Aliases": [ + "ord" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaximumNumberOfInverts", + "Type": "Int", + "Desc": "Limit the number of keys used to generate the slot name to this many. 0 means no invert hashing, -1 means no limit.", + "Aliases": [ + "ih" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:numberOfBits:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "Number of bits to hash into. Must be between 1 and 30, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 16 + }, + { + "Name": "OutputKind", + "Type": { + "Kind": "Enum", + "Values": [ + "Bag", + "Indicator", + "Key", + "Binary" + ] + }, + "Desc": "Output kind: Bag (multi-set vector), Ind (indicator vector), or Key (index)", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": "Bag" + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + }, + { + "Name": "Ordered", + "Type": "Bool", + "Desc": "Whether the position of each term should be included in the hash", + "Aliases": [ + "ord" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaximumNumberOfInverts", + "Type": "Int", + "Desc": "Limit the number of keys used to generate the slot name to this many. 0 means no invert hashing, -1 means no limit.", + "Aliases": [ + "ih" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.CategoricalOneHotVectorizer", + "Desc": "Converts the categorical value into an indicator array by building a dictionary of categories based on the data and using the id in the dictionary as the index in the array.", + "FriendlyName": "Categorical Transform", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "OutputKind", + "Type": { + "Kind": "Enum", + "Values": [ + "Bag", + "Indicator", + "Key", + "Binary" + ] + }, + "Desc": "Output kind: Bag (multi-set vector), Ind (indicator vector), Key (index), or Binary encoded indicator vector", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of terms to keep when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of keys to keep per column when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "OutputKind", + "Type": { + "Kind": "Enum", + "Values": [ + "Bag", + "Indicator", + "Key", + "Binary" + ] + }, + "Desc": "Output kind: Bag (multi-set vector), Ind (indicator vector), or Key (index)", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": "Indicator" + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 113.0, + "IsNullable": false, + "Default": "ByOccurrence" + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 114.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.CharacterTokenizer", + "Desc": "Character-oriented tokenizer where text is considered a sequence of characters.", + "FriendlyName": "Character Tokenizer Transform", + "ShortName": "CharToken", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseMarkerChars", + "Type": "Bool", + "Desc": "Whether to mark the beginning/end of each row/slot with start of text character (0x02)/end of text character (0x03)", + "Aliases": [ + "mark" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ColumnConcatenator", + "Desc": "Concatenates one or more columns of the same item type.", + "FriendlyName": "Concat Transform", + "ShortName": "Concat", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:srcs)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ColumnCopier", + "Desc": "Duplicates columns from the dataset", + "FriendlyName": "Copy Columns Transform", + "ShortName": "Copy", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ColumnSelector", + "Desc": "Selects a set of columns, dropping all others", + "FriendlyName": "Select Columns", + "ShortName": null, + "Inputs": [ + { + "Name": "KeepColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of columns to keep.", + "Aliases": [ + "keepcol" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "DropColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of columns to drop.", + "Aliases": [ + "dropcol" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "KeepHidden", + "Type": "Bool", + "Desc": "Specifies whether to keep or remove hidden columns.", + "Aliases": [ + "hidden" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "IgnoreMissing", + "Type": "Bool", + "Desc": "Specifies whether to ignore columns that are missing from the input.", + "Aliases": [ + "ignore" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ColumnTypeConverter", + "Desc": "Converts a column to a different type, using standard conversions.", + "FriendlyName": "Convert Transform", + "ShortName": "Convert", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "ResultType", + "Type": { + "Kind": "Enum", + "Values": [ + "I1", + "U1", + "I2", + "U2", + "I4", + "U4", + "I8", + "U8", + "R4", + "Num", + "R8", + "TX", + "Text", + "TXT", + "BL", + "Bool", + "TimeSpan", + "TS", + "DT", + "DateTime", + "DZ", + "DateTimeZone", + "UG", + "U16" + ] + }, + "Desc": "The result type", + "Aliases": [ + "type" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Range", + "Type": "String", + "Desc": "For a key column, this defines the range of values", + "Aliases": [ + "key" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:type:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ResultType", + "Type": { + "Kind": "Enum", + "Values": [ + "I1", + "U1", + "I2", + "U2", + "I4", + "U4", + "I8", + "U8", + "R4", + "Num", + "R8", + "TX", + "Text", + "TXT", + "BL", + "Bool", + "TimeSpan", + "TS", + "DT", + "DateTime", + "DZ", + "DateTimeZone", + "UG", + "U16" + ] + }, + "Desc": "The result type", + "Aliases": [ + "type" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Range", + "Type": "String", + "Desc": "For a key column, this defines the range of values", + "Aliases": [ + "key" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.CombinerByContiguousGroupId", + "Desc": "Groups values of a scalar column into a vector, by a contiguous group ID", + "FriendlyName": "Group Transform", + "ShortName": "Group", + "Inputs": [ + { + "Name": "GroupKey", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Columns to group by", + "Aliases": [ + "g" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Columns to group together", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ConditionalNormalizer", + "Desc": "Normalize the columns only if needed", + "FriendlyName": "Normalize If Needed", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ] + }, + { + "Name": "Transforms.CountTableBuilder", + "Desc": "Transforms the categorical column into the set of features: count of each label class, log-odds for each label class, back-off indicator. The input columns must be keys.", + "FriendlyName": "Count Table Transform", + "ShortName": "Count", + "Inputs": [ + { + "Name": "Columns", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "CountTable", + "Type": { + "Kind": "Component", + "ComponentKind": "CountTableBuilder" + }, + "Desc": "Count table settings", + "Aliases": [ + "table" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PriorCoefficient", + "Type": "Float", + "Desc": "The coefficient with which to apply the prior smoothing to the features", + "Aliases": [ + "prior" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LaplaceScale", + "Type": "Float", + "Desc": "Laplacian noise diversity/scale-parameter. Suggest keeping it less than 1.", + "Aliases": [ + "laplace" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Seed for the random generator for the laplacian noise.", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "CountTable", + "Type": { + "Kind": "Component", + "ComponentKind": "CountTableBuilder" + }, + "Desc": "Count table settings", + "Aliases": [ + "table" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "CMSketch" + } + }, + { + "Name": "PriorCoefficient", + "Type": "Float", + "Desc": "The coefficient with which to apply the prior smoothing to the features", + "Aliases": [ + "prior" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "LaplaceScale", + "Type": "Float", + "Desc": "Laplacian noise diversity/scale-parameter. Suggest keeping it less than 1.", + "Aliases": [ + "laplace" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Seed for the random generator for the laplacian noise.", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Label column", + "Aliases": [ + "label", + "lab" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + }, + { + "Name": "InitialCountsModel", + "Type": "String", + "Desc": "Optional model file to load counts from. If this is specified all other options are ignored.", + "Aliases": [ + "inmodel", + "extfile" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "SharedTable", + "Type": "Bool", + "Desc": "Keep counts for all columns in one shared count table", + "Aliases": [ + "shared" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.CountTargetEncoder", + "Desc": "Transforms the categorical column into the set of features: count of each label class, log-odds for each label class, back-off indicator. The columns can be of arbitrary type.", + "FriendlyName": "Count Table Transform", + "ShortName": "Count", + "Inputs": [ + { + "Name": "Columns", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "CountTable", + "Type": { + "Kind": "Component", + "ComponentKind": "CountTableBuilder" + }, + "Desc": "Count table settings", + "Aliases": [ + "table" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PriorCoefficient", + "Type": "Float", + "Desc": "The coefficient with which to apply the prior smoothing to the features", + "Aliases": [ + "prior" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LaplaceScale", + "Type": "Float", + "Desc": "Laplacian noise diversity/scale-parameter. Suggest keeping it less than 1.", + "Aliases": [ + "laplace" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Seed for the random generator for the laplacian noise.", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Combine", + "Type": "Bool", + "Desc": "Whether the values need to be combined for a single hash", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Combine", + "Type": "Bool", + "Desc": "Whether the values need to be combined for a single hash", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CountTable", + "Type": { + "Kind": "Component", + "ComponentKind": "CountTableBuilder" + }, + "Desc": "Count table settings", + "Aliases": [ + "table" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "CMSketch" + } + }, + { + "Name": "PriorCoefficient", + "Type": "Float", + "Desc": "The coefficient with which to apply the prior smoothing to the features", + "Aliases": [ + "prior" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "LaplaceScale", + "Type": "Float", + "Desc": "Laplacian noise diversity/scale-parameter. Suggest keeping it less than 1.", + "Aliases": [ + "laplace" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "Seed for the random generator for the laplacian noise.", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "Label column", + "Aliases": [ + "label", + "lab" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + }, + { + "Name": "InitialCountsModel", + "Type": "String", + "Desc": "Optional model file to load counts from. If this is specified all other options are ignored.", + "Aliases": [ + "inmodel", + "extfile" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "SharedTable", + "Type": "Bool", + "Desc": "Keep counts for all columns in one shared count table", + "Aliases": [ + "shared" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "Number of bits to hash into. Must be between 1 and 31, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 31 + }, + { + "Name": "HashingSeed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.DatasetScorer", + "Desc": "Score a dataset with a predictor model", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The dataset to be scored", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model to apply to data", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "Suffix", + "Type": "String", + "Desc": "Suffix to append to the score columns", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "ScoredData", + "Type": "DataView", + "Desc": "The scored dataset" + }, + { + "Name": "ScoringTransform", + "Type": "TransformModel", + "Desc": "The scoring transform" + } + ] + }, + { + "Name": "Transforms.DatasetTransformScorer", + "Desc": "Score a dataset with a transform model", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "The dataset to be scored", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "TransformModel", + "Type": "TransformModel", + "Desc": "The transform model to apply to data", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "ScoredData", + "Type": "DataView", + "Desc": "The scored dataset" + }, + { + "Name": "ScoringTransform", + "Type": "TransformModel", + "Desc": "The scoring transform" + } + ] + }, + { + "Name": "Transforms.Dictionarizer", + "Desc": "Converts input values (words, numbers, etc.) to index in a dictionary.", + "FriendlyName": "Term Transform", + "ShortName": "TermTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of terms to keep when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of keys to keep per column when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 113.0, + "IsNullable": false, + "Default": "ByOccurrence" + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 114.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.FeatureCombiner", + "Desc": "Combines all the features into one feature column.", + "FriendlyName": "Feature Combiner", + "ShortName": "fc", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Features", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Features", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.FeatureContributionCalculationTransformer", + "Desc": "For each data point, calculates the contribution of individual features to the model prediction.", + "FriendlyName": "Feature Contribution Calculation", + "ShortName": null, + "Inputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model to apply to data", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FeatureColumn", + "Type": "String", + "Desc": "Name of feature column", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "Top", + "Type": "Int", + "Desc": "Number of top contributions", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "Bottom", + "Type": "Int", + "Desc": "Number of bottom contributions", + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "Normalize", + "Type": "Bool", + "Desc": "Whether or not output of Features contribution should be normalized", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.FeatureSelectorByCount", + "Desc": "Selects the slots for which the count of non-default values is greater than or equal to a threshold.", + "FriendlyName": "Count Feature Selection Transform", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Columns to use for feature selection", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Count", + "Type": "Int", + "Desc": "If the count of non-default values for a slot is greater than or equal to this threshold, the slot is preserved", + "Aliases": [ + "c" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.FeatureSelectorByMutualInformation", + "Desc": "Selects the top k slots across all specified columns ordered by their mutual information with the label column.", + "FriendlyName": "Mutual Information Feature Selection Transform", + "ShortName": "MIFeatureSelection", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Columns to use for feature selection", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "SlotsInOutput", + "Type": "Int", + "Desc": "The maximum number of slots to preserve in output", + "Aliases": [ + "topk", + "numSlotsToKeep" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 1000 + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "NumBins", + "Type": "Int", + "Desc": "Max number of bins for R4/R8 columns, power of 2 recommended", + "Aliases": [ + "bins" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 256 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.GlobalContrastNormalizer", + "Desc": "Performs a global contrast normalization on input values: Y = (s * X - M) / D, where s is a scale, M is mean and D is either L2 norm or standard deviation.", + "FriendlyName": "Global Contrast Normalization Transform", + "ShortName": "Gcn", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "UseStdDev", + "Type": "Bool", + "Desc": "Normalize by standard deviation rather than L2 norm", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale features by this value", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "SubMean", + "Type": "Bool", + "Desc": "Subtract mean from each value before normalizing", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "SubMean", + "Type": "Bool", + "Desc": "Subtract mean from each value before normalizing", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseStdDev", + "Type": "Bool", + "Desc": "Normalize by standard deviation rather than L2 norm", + "Aliases": [ + "useStd" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale features by this value", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.HashConverter", + "Desc": "Converts column values into hashes. This transform accepts both numeric and text inputs, both single and vector-valued columns. ", + "FriendlyName": "Hash Join Transform", + "ShortName": "HashJoin", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Join", + "Type": "Bool", + "Desc": "Whether the values need to be combined for a single hash", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "CustomSlotMap", + "Type": "String", + "Desc": "Which slots should be combined together. Example: 0,3,5;0,1;3;2,1,0. Overrides 'join'.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "Number of bits to hash into. Must be between 1 and 31, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Ordered", + "Type": "Bool", + "Desc": "Whether the position of each term should be included in the hash", + "Aliases": [ + "ord" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "Number of bits to hash into. Must be between 1 and 31, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 31 + }, + { + "Name": "Join", + "Type": "Bool", + "Desc": "Whether the values need to be combined for a single hash", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + }, + { + "Name": "Ordered", + "Type": "Bool", + "Desc": "Whether the position of each term should be included in the hash", + "Aliases": [ + "ord" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ImageGrayscale", + "Desc": "Convert image into grayscale.", + "FriendlyName": "Image Greyscale Transform", + "ShortName": "ImageGrayscaleTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ImageLoader", + "Desc": "Load images from files.", + "FriendlyName": "Image Loader Transform", + "ShortName": "ImageLoaderTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ImageFolder", + "Type": "String", + "Desc": "Folder where to search for images", + "Aliases": [ + "folder" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ImagePixelExtractor", + "Desc": "Extract color plane(s) from an image. Options include scaling, offset and conversion to floating point.", + "FriendlyName": "Image Pixel Extractor Transform", + "ShortName": "ImagePixelExtractor", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "UseAlpha", + "Type": "Bool", + "Desc": "Whether to use alpha channel", + "Aliases": [ + "alpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "UseRed", + "Type": "Bool", + "Desc": "Whether to use red channel", + "Aliases": [ + "red" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "UseGreen", + "Type": "Bool", + "Desc": "Whether to use green channel", + "Aliases": [ + "green" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "UseBlue", + "Type": "Bool", + "Desc": "Whether to use blue channel", + "Aliases": [ + "blue" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Order", + "Type": { + "Kind": "Enum", + "Values": [ + "ARGB", + "ARBG", + "ABRG", + "ABGR", + "AGRB", + "AGBR" + ] + }, + "Desc": "Order of channels", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Interleave", + "Type": "Bool", + "Desc": "Whether to separate each channel or interleave in specified order", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Convert", + "Type": "Bool", + "Desc": "Whether to convert to floating point", + "Aliases": [ + "conv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "Offset (pre-scale)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale factor", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseAlpha", + "Type": "Bool", + "Desc": "Whether to use alpha channel", + "Aliases": [ + "alpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseRed", + "Type": "Bool", + "Desc": "Whether to use red channel", + "Aliases": [ + "red" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "UseGreen", + "Type": "Bool", + "Desc": "Whether to use green channel", + "Aliases": [ + "green" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "UseBlue", + "Type": "Bool", + "Desc": "Whether to use blue channel", + "Aliases": [ + "blue" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Order", + "Type": { + "Kind": "Enum", + "Values": [ + "ARGB", + "ARBG", + "ABRG", + "ABGR", + "AGRB", + "AGBR" + ] + }, + "Desc": "Order of colors.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "ARGB" + }, + { + "Name": "Interleave", + "Type": "Bool", + "Desc": "Whether to separate each channel or interleave in specified order", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Convert", + "Type": "Bool", + "Desc": "Whether to convert to floating point", + "Aliases": [ + "conv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "Offset (pre-scale)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale factor", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ImageResizer", + "Desc": "Scales an image to specified dimensions using one of the three scale types: isotropic with padding, isotropic with cropping or anisotropic. In case of isotropic padding, transparent color is used to pad resulting image.", + "FriendlyName": "Image Resizer Transform", + "ShortName": "ImageScalerTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "ImageWidth", + "Type": "Int", + "Desc": "Width of the resized image", + "Aliases": [ + "width" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ImageHeight", + "Type": "Int", + "Desc": "Height of the resized image", + "Aliases": [ + "height" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Resizing", + "Type": { + "Kind": "Enum", + "Values": [ + "IsoPad", + "IsoCrop", + "Fill" + ] + }, + "Desc": "Resizing method", + "Aliases": [ + "scale" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "CropAnchor", + "Type": { + "Kind": "Enum", + "Values": [ + "Right", + "Left", + "Top", + "Bottom", + "Center" + ] + }, + "Desc": "Anchor for cropping", + "Aliases": [ + "anchor" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ImageWidth", + "Type": "Int", + "Desc": "Resized width of the image", + "Aliases": [ + "width" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "ImageHeight", + "Type": "Int", + "Desc": "Resized height of the image", + "Aliases": [ + "height" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Resizing", + "Type": { + "Kind": "Enum", + "Values": [ + "IsoPad", + "IsoCrop", + "Fill" + ] + }, + "Desc": "Resizing method", + "Aliases": [ + "scale" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "IsoCrop" + }, + { + "Name": "CropAnchor", + "Type": { + "Kind": "Enum", + "Values": [ + "Right", + "Left", + "Top", + "Bottom", + "Center" + ] + }, + "Desc": "Anchor for cropping", + "Aliases": [ + "anchor" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Center" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.KeyToTextConverter", + "Desc": "KeyToValueTransform utilizes KeyValues metadata to map key indices to the corresponding values in the KeyValues metadata.", + "FriendlyName": "Key To Value Transform", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LabelColumnKeyBooleanConverter", + "Desc": "Transforms the label to either key or bool (if needed) to make it suitable for classification.", + "FriendlyName": "Prepare Classification Label", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "The label column", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Convert the key values to text", + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LabelIndicator", + "Desc": "Label remapper used by OVA", + "FriendlyName": "LabelIndicator", + "ShortName": "LabelIndictator", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "ClassIndex", + "Type": "Int", + "Desc": "The positive example class for binary classification.", + "Aliases": [ + "index" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ClassIndex", + "Type": "Int", + "Desc": "Label of the positive class.", + "Aliases": [ + "index" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LabelToFloatConverter", + "Desc": "Transforms the label to float to make it suitable for regression.", + "FriendlyName": "Prepare Regression Label", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "LabelColumn", + "Type": "String", + "Desc": "The label column", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LightLda", + "Desc": "The LDA transform implements LightLDA, a state-of-the-art implementation of Latent Dirichlet Allocation.", + "FriendlyName": "Latent Dirichlet Allocation Transform", + "ShortName": "LightLda", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "NumTopic", + "Type": "Int", + "Desc": "The number of topics", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "AlphaSum", + "Type": "Float", + "Desc": "Dirichlet prior on document-topic vectors", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Beta", + "Type": "Float", + "Desc": "Dirichlet prior on vocab-topic vectors", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Mhstep", + "Type": "Int", + "Desc": "Number of Metropolis Hasting step", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "LikelihoodInterval", + "Type": "Int", + "Desc": "Compute log likelihood over local dataset on this iteration interval", + "Aliases": [ + "llInterval" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumThreads", + "Type": "Int", + "Desc": "The number of training threads", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumMaxDocToken", + "Type": "Int", + "Desc": "The threshold of maximum count of tokens per doc", + "Aliases": [ + "maxNumToken" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumSummaryTermPerTopic", + "Type": "Int", + "Desc": "The number of words to summarize the topic", + "Aliases": [ + "ns" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "NumBurninIterations", + "Type": "Int", + "Desc": "The number of burn-in iterations", + "Aliases": [ + "burninIter" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": 10 + }, + { + "Name": "ResetRandomGenerator", + "Type": "Bool", + "Desc": "Reset the random number generator for each document", + "Aliases": [ + "reset" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "OutputTopicWordSummary", + "Type": "Bool", + "Desc": "Whether to output the topic-word summary in text format when saving the model to disk", + "Aliases": [ + "summary" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:srcs)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 49.0, + "IsNullable": false + }, + { + "Name": "NumTopic", + "Type": "Int", + "Desc": "The number of topics", + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 40, + 100, + 200 + ] + } + }, + { + "Name": "NumThreads", + "Type": "Int", + "Desc": "The number of training threads. Default value depends on number of logical processors.", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "NumMaxDocToken", + "Type": "Int", + "Desc": "The threshold of maximum count of tokens per doc", + "Aliases": [ + "maxNumToken" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 512 + }, + { + "Name": "AlphaSum", + "Type": "Float", + "Desc": "Dirichlet prior on document-topic vectors", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 100, + 200 + ] + } + }, + { + "Name": "Beta", + "Type": "Float", + "Desc": "Dirichlet prior on vocab-topic vectors", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.01, + 0.015, + 0.07, + 0.02 + ] + } + }, + { + "Name": "Mhstep", + "Type": "Int", + "Desc": "Number of Metropolis Hasting step", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 4, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 2, + 4, + 8, + 16 + ] + } + }, + { + "Name": "NumIterations", + "Type": "Int", + "Desc": "Number of iterations", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 200, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 100, + 200, + 300, + 400 + ] + } + }, + { + "Name": "LikelihoodInterval", + "Type": "Int", + "Desc": "Compute log likelihood over local dataset on this iteration interval", + "Aliases": [ + "llInterval" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "NumSummaryTermPerTopic", + "Type": "Int", + "Desc": "The number of words to summarize the topic", + "Aliases": [ + "ns" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10 + }, + { + "Name": "NumBurninIterations", + "Type": "Int", + "Desc": "The number of burn-in iterations", + "Aliases": [ + "burninIter" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 10, + 20, + 30, + 40 + ] + } + }, + { + "Name": "ResetRandomGenerator", + "Type": "Bool", + "Desc": "Reset the random number generator for each document", + "Aliases": [ + "reset" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "OutputTopicWordSummary", + "Type": "Bool", + "Desc": "Whether to output the topic-word summary in text format when saving the model to disk", + "Aliases": [ + "summary" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LogMeanVarianceNormalizer", + "Desc": "Normalizes the data based on the computed mean and variance of the logarithm of the data.", + "FriendlyName": "LogMeanVar Normalizer", + "ShortName": "LogMeanVar", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseCdf", + "Type": "Bool", + "Desc": "Whether to use CDF as the output", + "Aliases": [ + "cdf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.LpNormalizer", + "Desc": "Normalize vectors (rows) individually by rescaling them to unit norm (L2, L1 or LInf). Performs the following operation on a vector X: Y = (X - M) / D, where M is mean and D is either L2 norm, L1 norm or LInf norm.", + "FriendlyName": "Lp-Norm Normalizer", + "ShortName": "lpnorm", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Norm", + "Type": { + "Kind": "Enum", + "Values": [ + "L2", + "StandardDeviation", + "L1", + "Infinity" + ] + }, + "Desc": "The norm to use to normalize each sample", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "SubMean", + "Type": "Bool", + "Desc": "Subtract mean from each value before normalizing", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Norm", + "Type": { + "Kind": "Enum", + "Values": [ + "L2", + "StandardDeviation", + "L1", + "Infinity" + ] + }, + "Desc": "The norm to use to normalize each sample", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": "L2" + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "SubMean", + "Type": "Bool", + "Desc": "Subtract mean from each value before normalizing", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ManyHeterogeneousModelCombiner", + "Desc": "Combines a sequence of TransformModels and a PredictorModel into a single PredictorModel.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "TransformModels", + "Type": { + "Kind": "Array", + "ItemType": "TransformModel" + }, + "Desc": "Transform model", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Predictor model", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Predictor model" + } + ] + }, + { + "Name": "Transforms.MeanVarianceNormalizer", + "Desc": "Normalizes the data based on the computed mean and variance of the data.", + "FriendlyName": "MeanVar Normalizer", + "ShortName": "MeanVar", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseCdf", + "Type": "Bool", + "Desc": "Whether to use CDF as the output", + "Aliases": [ + "cdf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MinMaxNormalizer", + "Desc": "Normalizes the data based on the observed minimum and maximum values of the data.", + "FriendlyName": "Min-Max Normalizer", + "ShortName": "MinMax", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MissingValueHandler", + "Desc": "Handle missing values by replacing them with either the default value or the mean/min/max value (for non-text columns only). An indicator column can optionally be concatenated, if the input column type is numeric.", + "FriendlyName": "NA Handle Transform", + "ShortName": "NAHandle", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Kind", + "Type": { + "Kind": "Enum", + "Values": [ + "DefaultValue", + "Mean", + "Minimum", + "Maximum" + ] + }, + "Desc": "The replacement method to utilize", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ImputeBySlot", + "Type": "Bool", + "Desc": "Whether to impute values by slot", + "Aliases": [ + "slot" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ConcatIndicator", + "Type": "Bool", + "Desc": "Whether or not to concatenate an indicator vector column to the value column", + "Aliases": [ + "ind" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:rep:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ReplaceWith", + "Type": { + "Kind": "Enum", + "Values": [ + "DefaultValue", + "Mean", + "Minimum", + "Maximum" + ] + }, + "Desc": "The replacement method to utilize", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Def" + }, + { + "Name": "ImputeBySlot", + "Type": "Bool", + "Desc": "Whether to impute values by slot", + "Aliases": [ + "slot" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Concat", + "Type": "Bool", + "Desc": "Whether or not to concatenate an indicator vector column to the value column", + "Aliases": [ + "ind" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MissingValueIndicator", + "Desc": "Create a boolean output column with the same number of slots as the input column, where the output value is true if the value in the input column is missing.", + "FriendlyName": "NA Indicator Transform", + "ShortName": "NAInd", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MissingValuesDropper", + "Desc": "Removes NAs from vector columns.", + "FriendlyName": "NA Drop Transform", + "ShortName": "NADrop", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "Columns to drop the NAs for", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MissingValuesRowDropper", + "Desc": "Filters out rows that contain missing values.", + "FriendlyName": "NA Filter", + "ShortName": "NAFilter", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Column", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Complement", + "Type": "Bool", + "Desc": "If true, keep only rows that contain NA values, and filter the rest.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.MissingValueSubstitutor", + "Desc": "Create an output column of the same type and size of the input column, where missing values are replaced with either the default value or the mean/min/max value (for non-text columns only).", + "FriendlyName": "NA Replace Transform", + "ShortName": "NARep", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "ReplacementString", + "Type": "String", + "Desc": "Replacement value for NAs (uses default value if not given)", + "Aliases": [ + "rep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Kind", + "Type": { + "Kind": "Enum", + "Values": [ + "DefaultValue", + "Mean", + "Minimum", + "Maximum", + "SpecifiedValue", + "Mode" + ] + }, + "Desc": "The replacement method to utilize", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Slot", + "Type": "Bool", + "Desc": "Whether to impute values by slot", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:rep:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ReplacementKind", + "Type": { + "Kind": "Enum", + "Values": [ + "DefaultValue", + "Mean", + "Minimum", + "Maximum", + "SpecifiedValue", + "Mode" + ] + }, + "Desc": "The replacement method to utilize", + "Aliases": [ + "kind" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Def" + }, + { + "Name": "ImputeBySlot", + "Type": "Bool", + "Desc": "Whether to impute values by slot", + "Aliases": [ + "slot" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ModelCombiner", + "Desc": "Combines a sequence of TransformModels into a single model", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "Models", + "Type": { + "Kind": "Array", + "ItemType": "TransformModel" + }, + "Desc": "Input models", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputModel", + "Type": "TransformModel", + "Desc": "Combined model" + } + ] + }, + { + "Name": "Transforms.NGramTranslator", + "Desc": "Produces a bag of counts of n-grams (sequences of consecutive values of length 1-n) in a given vector of keys. It does so by building a dictionary of n-grams and using the id in the dictionary as the index in the bag.", + "FriendlyName": "NGram Transform", + "ShortName": "NgramTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "NgramLength", + "Type": "Int", + "Desc": "Maximum n-gram length", + "Aliases": [ + "ngram" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "AllLengths", + "Type": "Bool", + "Desc": "Whether to include all n-gram lengths up to NgramLength or only NgramLength", + "Aliases": [ + "all" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "SkipLength", + "Type": "Int", + "Desc": "Maximum number of tokens to skip when constructing an n-gram", + "Aliases": [ + "skips" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxNumTerms", + "Type": { + "Kind": "Array", + "ItemType": "Int" + }, + "Desc": "Maximum number of n-grams to store in the dictionary", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Weighting", + "Type": { + "Kind": "Enum", + "Values": [ + "Tf", + "Idf", + "TfIdf" + ] + }, + "Desc": "Statistical measure used to evaluate how important a word is to a document in a corpus", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NgramLength", + "Type": "Int", + "Desc": "Maximum n-gram length", + "Aliases": [ + "ngram" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2 + }, + { + "Name": "AllLengths", + "Type": "Bool", + "Desc": "Whether to store all n-gram lengths up to ngramLength, or only ngramLength", + "Aliases": [ + "all" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "SkipLength", + "Type": "Int", + "Desc": "Maximum number of tokens to skip when constructing an n-gram", + "Aliases": [ + "skips" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MaxNumTerms", + "Type": { + "Kind": "Array", + "ItemType": "Int" + }, + "Desc": "Maximum number of n-grams to store in the dictionary", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + 10000000 + ] + }, + { + "Name": "Weighting", + "Type": { + "Kind": "Enum", + "Values": [ + "Tf", + "Idf", + "TfIdf" + ] + }, + "Desc": "The weighting criteria", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Tf" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.NoOperation", + "Desc": "Does nothing.", + "FriendlyName": "No Op", + "ShortName": "Nop", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.OptionalColumnCreator", + "Desc": "If the source column does not exist after deserialization, create a column with the right type and default values.", + "FriendlyName": "Optional Column Transform", + "ShortName": "optional", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "New column definition(s)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.PcaCalculator", + "Desc": "PCA is a dimensionality-reduction transform which computes the projection of a numeric vector onto a low-rank subspace.", + "FriendlyName": "Principal Component Analysis Transform", + "ShortName": "Pca", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "WeightColumn", + "Type": "String", + "Desc": "The name of the weight column", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Rank", + "Type": "Int", + "Desc": "The number of components in the PCA", + "Aliases": [ + "k" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Oversampling", + "Type": "Int", + "Desc": "Oversampling parameter for randomized PCA training", + "Aliases": [ + "over" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Center", + "Type": "Bool", + "Desc": "If enabled, data is centered to be zero mean", + "Aliases": [ + "center" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed for random number generation", + "Aliases": [ + "seed" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "The name of the weight column", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Rank", + "Type": "Int", + "Desc": "The number of components in the PCA", + "Aliases": [ + "k" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 20 + }, + { + "Name": "Oversampling", + "Type": "Int", + "Desc": "Oversampling parameter for randomized PCA training", + "Aliases": [ + "over" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 20 + }, + { + "Name": "Center", + "Type": "Bool", + "Desc": "If enabled, data is centered to be zero mean", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed for random number generation", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.PermutationFeatureImportance", + "Desc": "Permutation Feature Importance (PFI)", + "FriendlyName": "PFI", + "ShortName": "PFI", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The path to the model file", + "Aliases": [ + "path" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + }, + { + "Name": "UseFeatureWeightFilter", + "Type": "Bool", + "Desc": "Use feature weights to pre-filter features", + "Aliases": [ + "usefw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NumberOfExamplesToUse", + "Type": "Int", + "Desc": "Limit the number of examples to evaluate on", + "Aliases": [ + "numexamples" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "PermutationCount", + "Type": "Int", + "Desc": "The number of permutations to perform", + "Aliases": [ + "permutations" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + } + ], + "Outputs": [ + { + "Name": "Metrics", + "Type": "DataView", + "Desc": "The PFI metrics" + } + ], + "InputKind": [ + "ITransformInput" + ] + }, + { + "Name": "Transforms.PredictedLabelColumnOriginalValueConverter", + "Desc": "Transforms a predicted label column to its original values, unless it is of type bool.", + "FriendlyName": "Convert Predicted Label", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictedLabelColumn", + "Type": "String", + "Desc": "The predicted label column", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RandomNumberGenerator", + "Desc": "Adds a column with a generated number sequence.", + "FriendlyName": "Generate Number Transform", + "ShortName": "Generate", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "UseCounter", + "Type": "Bool", + "Desc": "Use an auto-incremented integer starting at zero instead of a random number", + "Aliases": [ + "cnt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "The random seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:seed)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "UseCounter", + "Type": "Bool", + "Desc": "Use an auto-incremented integer starting at zero instead of a random number", + "Aliases": [ + "cnt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "The random seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 42 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RobustScalingNormalizer", + "Desc": "Optionally centers the data and scales based on the range of data and the quantile min and max values provided. This method is more robust to outliers.", + "FriendlyName": "Robust Scaling Normalizer", + "ShortName": "RobScal", + "Inputs": [ + { + "Name": "CenterData", + "Type": "Bool", + "Desc": "Should the data be centered around 0", + "Aliases": [ + "center" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "QuantileMin", + "Type": "UInt", + "Desc": "Minimum quantile value. Defaults to 25", + "Aliases": [ + "qmin" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 25 + }, + { + "Name": "QuantileMax", + "Type": "UInt", + "Desc": "Maximum quantile value. Defaults to 75", + "Aliases": [ + "qmax" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 75 + }, + { + "Name": "FixZero", + "Type": "Bool", + "Desc": "Whether to map zero to zero, preserving sparsity", + "Aliases": [ + "zero" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxTrainingExamples", + "Type": "Int", + "Desc": "Max number of examples used to train the normalizer", + "Aliases": [ + "maxtrain" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1000000000 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RowRangeFilter", + "Desc": "Filters a dataview on a column of type Single, Double or Key (contiguous). Keeps the values that are in the specified min/max range. NaNs are always filtered out. If the input is a Key type, the min/max are considered percentages of the number of values.", + "FriendlyName": "Range Filter", + "ShortName": "RangeFilter", + "Inputs": [ + { + "Name": "Column", + "Type": "String", + "Desc": "Column", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Min", + "Type": "Float", + "Desc": "Minimum value (0 to 1 for key types)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Max", + "Type": "Float", + "Desc": "Maximum value (0 to 1 for key types)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Complement", + "Type": "Bool", + "Desc": "If true, keep the values that fall outside the range.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "IncludeMin", + "Type": "Bool", + "Desc": "If true, include in the range the values that are equal to min.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "IncludeMax", + "Type": "Bool", + "Desc": "If true, include in the range the values that are equal to max.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RowSkipAndTakeFilter", + "Desc": "Allows limiting input to a subset of rows at an optional offset. Can be used to implement data paging.", + "FriendlyName": "Skip and Take Filter", + "ShortName": "SkipTake", + "Inputs": [ + { + "Name": "Skip", + "Type": "Int", + "Desc": "Number of items to skip", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Take", + "Type": "Int", + "Desc": "Number of items to take", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": true, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RowSkipFilter", + "Desc": "Allows limiting input to a subset of rows by skipping a number of rows.", + "FriendlyName": "Skip Filter", + "ShortName": "Skip", + "Inputs": [ + { + "Name": "Count", + "Type": "Int", + "Desc": "Number of items to skip", + "Aliases": [ + "c", + "n", + "s" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.RowTakeFilter", + "Desc": "Allows limiting input to a subset of rows by taking N first rows.", + "FriendlyName": "Take Filter", + "ShortName": "Take", + "Inputs": [ + { + "Name": "Count", + "Type": "Int", + "Desc": "Number of items to take", + "Aliases": [ + "c", + "n", + "t" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 9223372036854775807 + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.ScoreColumnSelector", + "Desc": "Selects only the last score columns and the extra columns specified in the arguments.", + "FriendlyName": "Choose Columns By Indices", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ExtraColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Extra columns to write", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.Scorer", + "Desc": "Turn the predictor model into a transform model", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "The predictor model to turn into a transform", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "ScoredData", + "Type": "DataView", + "Desc": "The scored dataset" + }, + { + "Name": "ScoringTransform", + "Type": "TransformModel", + "Desc": "The scoring transform" + } + ] + }, + { + "Name": "Transforms.Segregator", + "Desc": "Un-groups vector columns into sequences of rows, inverse of Group transform", + "FriendlyName": "Un-group Transform", + "ShortName": "Ungroup", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Columns to unroll, or 'pivot'", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + }, + { + "Name": "Mode", + "Type": { + "Kind": "Enum", + "Values": [ + "Inner", + "Outer", + "First" + ] + }, + "Desc": "Specifies how to unroll multiple pivot columns of different size.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Inner" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.SentimentAnalyzer", + "Desc": "Uses a pretrained sentiment model to score input strings", + "FriendlyName": "Sentiment Analyzing Transform", + "ShortName": "Senti", + "Inputs": [ + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column.", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column.", + "Aliases": [ + "dst" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.TensorFlowScorer", + "Desc": "Transforms the data using the TensorFlow model.", + "FriendlyName": "TensorFlowTransform", + "ShortName": "TFTransform", + "Inputs": [ + { + "Name": "ModelLocation", + "Type": "String", + "Desc": "TensorFlow model used by the transform. Please see https://www.tensorflow.org/mobile/prepare_models for more details.", + "Required": true, + "SortOrder": 0.0, + "IsNullable": false + }, + { + "Name": "InputColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "The names of the model inputs", + "Aliases": [ + "inputs" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "OutputColumns", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "The name of the outputs", + "Aliases": [ + "outputs" + ], + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + }, + { + "Name": "BatchSize", + "Type": "Int", + "Desc": "Number of samples to use for mini-batch training.", + "Required": false, + "SortOrder": 9.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "AddBatchDimensionInputs", + "Type": "Bool", + "Desc": "Add a batch dimension to the input e.g. input = [224, 224, 3] => [-1, 224, 224, 3].", + "Required": false, + "SortOrder": 16.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TreatOutputAsBatched", + "Type": "Bool", + "Desc": "If the first dimension of the output is unknown, should it be treated as batched or not. e.g. output = [-1] will be read as a vector of unknown length when this is false.", + "Required": false, + "SortOrder": 17.0, + "IsNullable": false, + "Default": true + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.TextFeaturizer", + "Desc": "A transform that turns a collection of text documents into numerical feature vectors. The feature vectors are normalized counts of (word and/or character) n-grams in a given tokenized text.", + "FriendlyName": "Text Transform", + "ShortName": "Text", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + }, + "Desc": "New column definition (optional form: name:srcs).", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Language", + "Type": { + "Kind": "Enum", + "Values": [ + "English", + "French", + "German", + "Dutch", + "Italian", + "Spanish", + "Japanese" + ] + }, + "Desc": "Dataset language or 'AutoDetect' to detect language per row.", + "Aliases": [ + "lang" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "English" + }, + { + "Name": "StopWordsRemover", + "Type": { + "Kind": "Component", + "ComponentKind": "StopWordsRemover" + }, + "Desc": "Stopwords remover.", + "Aliases": [ + "remover" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "TextCase", + "Type": { + "Kind": "Enum", + "Values": [ + "Lower", + "Upper", + "None" + ] + }, + "Desc": "Casing text using the rules of the invariant culture.", + "Aliases": [ + "case" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Lower" + }, + { + "Name": "KeepDiacritics", + "Type": "Bool", + "Desc": "Whether to keep diacritical marks or remove them.", + "Aliases": [ + "diac" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "KeepPunctuations", + "Type": "Bool", + "Desc": "Whether to keep punctuation marks or remove them.", + "Aliases": [ + "punc" + ], + "Required": false, + "SortOrder": 7.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "KeepNumbers", + "Type": "Bool", + "Desc": "Whether to keep numbers or remove them.", + "Aliases": [ + "num" + ], + "Required": false, + "SortOrder": 8.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "OutputTokensColumnName", + "Type": "String", + "Desc": "Column containing the transformed text tokens.", + "Aliases": [ + "tokens", + "showtext", + "showTransformedText" + ], + "Required": false, + "SortOrder": 9.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Dictionary", + "Type": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value, items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "ByOccurrence" + }, + { + "Name": "DropUnknowns", + "Type": "Bool", + "Desc": "Drop unknown terms instead of mapping them to NA term.", + "Aliases": [ + "dropna" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": false + } + ] + }, + "Desc": "A dictionary of allowed terms.", + "Aliases": [ + "dict" + ], + "Required": false, + "SortOrder": 10.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "WordFeatureExtractor", + "Type": { + "Kind": "Component", + "ComponentKind": "NgramExtractor" + }, + "Desc": "Ngram feature extractor to use for words (WordBag/WordHashBag).", + "Aliases": [ + "wordExtractor" + ], + "Required": false, + "SortOrder": 11.0, + "IsNullable": false, + "Default": { + "Name": "NGram", + "Settings": { + "NgramLength": 2, + "MaxNumTerms": [ + 10000000 + ] + } + } + }, + { + "Name": "CharFeatureExtractor", + "Type": { + "Kind": "Component", + "ComponentKind": "NgramExtractor" + }, + "Desc": "Ngram feature extractor to use for characters (WordBag/WordHashBag).", + "Aliases": [ + "charExtractor" + ], + "Required": false, + "SortOrder": 12.0, + "IsNullable": false, + "Default": { + "Name": "NGram", + "Settings": { + "NgramLength": 3, + "AllLengths": false, + "MaxNumTerms": [ + 10000000 + ] + } + } + }, + { + "Name": "VectorNormalizer", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "L1", + "L2", + "Infinity" + ] + }, + "Desc": "Normalize vectors (rows) individually by rescaling them to unit norm.", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 13.0, + "IsNullable": false, + "Default": "L2" + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.TextToKeyConverter", + "Desc": "Converts input values (words, numbers, etc.) to index in a dictionary.", + "FriendlyName": "To Key", + "ShortName": null, + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of terms to keep when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "MaxNumTerms", + "Type": "Int", + "Desc": "Maximum number of keys to keep per column when auto-training", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": 1000000 + }, + { + "Name": "Term", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of terms", + "Required": false, + "SortOrder": 106.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Sort", + "Type": { + "Kind": "Enum", + "Values": [ + "ByOccurrence", + "ByValue" + ] + }, + "Desc": "How items should be ordered when vectorized. By default, they will be in the order encountered. If by value items are sorted according to their default comparison, for example, text sorting will be case sensitive (for example, 'A' then 'Z' then 'a').", + "Required": false, + "SortOrder": 113.0, + "IsNullable": false, + "Default": "ByOccurrence" + }, + { + "Name": "TextKeyValues", + "Type": "Bool", + "Desc": "Whether key value metadata should be text, regardless of the actual input type", + "Aliases": [ + "textkv" + ], + "Required": false, + "SortOrder": 114.0, + "IsNullable": false, + "Default": false + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.TrainTestDatasetSplitter", + "Desc": "Split the dataset into train and test sets", + "FriendlyName": "Dataset Train-Test Split", + "ShortName": null, + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Fraction", + "Type": "Float", + "Desc": "Fraction of training data", + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 0.8 + }, + { + "Name": "StratificationColumn", + "Type": "String", + "Desc": "Stratification column", + "Aliases": [ + "strat" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "TrainData", + "Type": "DataView", + "Desc": "Training data" + }, + { + "Name": "TestData", + "Type": "DataView", + "Desc": "Testing data" + } + ] + }, + { + "Name": "Transforms.TreeLeafFeaturizer", + "Desc": "Trains a tree ensemble, or loads it from a file, then maps a numeric feature vector to three outputs: 1. A vector containing the individual tree outputs of the tree ensemble. 2. A vector indicating the leaves that the feature vector falls on in the tree ensemble. 3. A vector indicating the paths that the feature vector falls on in the tree ensemble. If a both a model file and a trainer are specified - will use the model file. If neither are specified, will train a default FastTree model. This can handle key labels by training a regression model towards their optionally permuted indices.", + "FriendlyName": "Tree Ensemble Featurization Transform", + "ShortName": "TreeFeat", + "Inputs": [ + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Trainer to use", + "Required": true, + "SortOrder": 10.0, + "IsNullable": false + }, + { + "Name": "Suffix", + "Type": "String", + "Desc": "Output column: The suffix to append to the default column names", + "Aliases": [ + "ex" + ], + "Required": false, + "SortOrder": 101.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "LabelPermutationSeed", + "Type": "Int", + "Desc": "If specified, determines the permutation seed for applying this featurizer to a multiclass problem.", + "Aliases": [ + "lps" + ], + "Required": false, + "SortOrder": 102.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "IFeaturizerInput", + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.TwoHeterogeneousModelCombiner", + "Desc": "Combines a TransformModel and a PredictorModel into a single PredictorModel.", + "FriendlyName": null, + "ShortName": null, + "Inputs": [ + { + "Name": "TransformModel", + "Type": "TransformModel", + "Desc": "Transform model", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Predictor model", + "Required": true, + "SortOrder": 2.0, + "IsNullable": false + } + ], + "Outputs": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel", + "Desc": "Predictor model" + } + ] + }, + { + "Name": "Transforms.VectorToImage", + "Desc": "Converts vector array into image type.", + "FriendlyName": "Vector To Image Transform", + "ShortName": "VectorToImageConverter", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "ContainsAlpha", + "Type": "Bool", + "Desc": "Whether to use alpha channel", + "Aliases": [ + "alpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ContainsRed", + "Type": "Bool", + "Desc": "Whether to use red channel", + "Aliases": [ + "red" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ContainsGreen", + "Type": "Bool", + "Desc": "Whether to use green channel", + "Aliases": [ + "green" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ContainsBlue", + "Type": "Bool", + "Desc": "Whether to use blue channel", + "Aliases": [ + "blue" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Order", + "Type": { + "Kind": "Enum", + "Values": [ + "ARGB", + "ARBG", + "ABRG", + "ABGR", + "AGRB", + "AGBR" + ] + }, + "Desc": "Order of channels", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Interleave", + "Type": "Bool", + "Desc": "Whether to separate each channel or interleave in specified order", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ImageWidth", + "Type": "Int", + "Desc": "Width of the image", + "Aliases": [ + "width" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "ImageHeight", + "Type": "Int", + "Desc": "Height of the image", + "Aliases": [ + "height" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "Offset (pre-scale)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale factor", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DefaultAlpha", + "Type": "Int", + "Desc": "Default value for alpha channel. Will be used if ContainsAlpha set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DefaultRed", + "Type": "Int", + "Desc": "Default value for red channel. Will be used if ContainsRed set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DefaultGreen", + "Type": "Int", + "Desc": "Default value for green channel. Will be used if ContainsGreen set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "DefaultBlue", + "Type": "Int", + "Desc": "Default value for blue channel. Will be used if ContainsGreen set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "ContainsAlpha", + "Type": "Bool", + "Desc": "Whether to use alpha channel", + "Aliases": [ + "alpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "ContainsRed", + "Type": "Bool", + "Desc": "Whether to use red channel", + "Aliases": [ + "red" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "ContainsGreen", + "Type": "Bool", + "Desc": "Whether to use green channel", + "Aliases": [ + "green" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "ContainsBlue", + "Type": "Bool", + "Desc": "Whether to use blue channel", + "Aliases": [ + "blue" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Order", + "Type": { + "Kind": "Enum", + "Values": [ + "ARGB", + "ARBG", + "ABRG", + "ABGR", + "AGRB", + "AGBR" + ] + }, + "Desc": "Order of colors.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "ARGB" + }, + { + "Name": "Interleave", + "Type": "Bool", + "Desc": "Whether to separate each channel or interleave in specified order", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "ImageWidth", + "Type": "Int", + "Desc": "Width of the image", + "Aliases": [ + "width" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "ImageHeight", + "Type": "Int", + "Desc": "Height of the image", + "Aliases": [ + "height" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "Offset (pre-scale)", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Scale", + "Type": "Float", + "Desc": "Scale factor", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "DefaultAlpha", + "Type": "Int", + "Desc": "Default value for alpha channel. Will be used if ContainsAlpha set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "DefaultRed", + "Type": "Int", + "Desc": "Default value for red channel. Will be used if ContainsRed set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "DefaultGreen", + "Type": "Int", + "Desc": "Default value for green channel. Will be used if ContainsGreen set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "DefaultBlue", + "Type": "Int", + "Desc": "Default value for blue channel. Will be used if ContainsBlue set to false", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.WordEmbeddings", + "Desc": "Word Embeddings transform is a text featurizer which converts vectors of text tokens into sentence vectors using a pre-trained model", + "FriendlyName": "Word Embeddings Transform", + "ShortName": "WordEmbeddings", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s) (optional form: name:src)", + "Aliases": [ + "col" + ], + "Required": true, + "SortOrder": 0.0, + "IsNullable": false + }, + { + "Name": "ModelKind", + "Type": { + "Kind": "Enum", + "Values": [ + "GloVe50D", + "GloVe100D", + "GloVe200D", + "GloVe300D", + "GloVeTwitter25D", + "GloVeTwitter50D", + "GloVeTwitter100D", + "GloVeTwitter200D", + "FastTextWikipedia300D", + "SentimentSpecificWordEmbedding" + ] + }, + "Desc": "Pre-trained model used to create the vocabulary", + "Aliases": [ + "model" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": true, + "Default": "SentimentSpecificWordEmbedding" + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "CustomLookupTable", + "Type": "String", + "Desc": "Filename for custom word embedding model", + "Aliases": [ + "dataFile" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + }, + { + "Name": "Transforms.WordTokenizer", + "Desc": "The input to this transform is text, and the output is a vector of text containing the words (tokens) in the original text. The separator is space, but can be specified as any other character (or multiple characters) if needed.", + "FriendlyName": "Tokenize Text Transform", + "ShortName": "TokenizeTextTransform", + "Inputs": [ + { + "Name": "Column", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "TermSeparators", + "Type": "String", + "Desc": "Comma separated set of term separator(s). Commonly: 'space', 'comma', 'semicolon' or other single character.", + "Aliases": [ + "sep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the new column", + "Aliases": [ + "name" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Source", + "Type": "String", + "Desc": "Name of the source column", + "Aliases": [ + "src" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + } + }, + "Desc": "New column definition(s)", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Data", + "Type": "DataView", + "Desc": "Input dataset", + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "CharArrayTermSeparators", + "Type": { + "Kind": "Array", + "ItemType": "Char" + }, + "Desc": "Array of single character term separator(s). By default uses space character separator.", + "Aliases": [ + "sep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ], + "Outputs": [ + { + "Name": "OutputData", + "Type": "DataView", + "Desc": "Transformed dataset" + }, + { + "Name": "Model", + "Type": "TransformModel", + "Desc": "Transform model" + } + ], + "InputKind": [ + "ITransformInput" + ], + "OutputKind": [ + "ITransformOutput" + ] + } + ], + "Components": [ + { + "Kind": "BoosterParameterFunction", + "Components": [ + { + "Name": "dart", + "Desc": "Dropouts meet Multiple Additive Regresion Trees. See https://arxiv.org/abs/1505.01866", + "FriendlyName": "Tree Dropout Tree Booster", + "Settings": [ + { + "Name": "TreeDropFraction", + "Type": "Float", + "Desc": "The drop ratio for trees. Range:(0,1).", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "MaximumNumberOfDroppedTreesPerRound", + "Type": "Int", + "Desc": "Maximum number of dropped trees in a boosting round.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1, + "Range": { + "Inf": 0, + "Max": 2147483647 + } + }, + { + "Name": "SkipDropFraction", + "Type": "Float", + "Desc": "Probability for not dropping in a boosting round.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.5, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "XgboostDartMode", + "Type": "Bool", + "Desc": "True will enable xgboost dart mode.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UniformDrop", + "Type": "Bool", + "Desc": "True will enable uniform drop.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MinimumSplitGain", + "Type": "Float", + "Desc": "Minimum loss reduction required to make a further partition on a leaf node of the tree. the larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "MaximumTreeDepth", + "Type": "Int", + "Desc": "Maximum depth of a tree. 0 means no limit. However, tree still grows by best-first.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "MinimumChildWeight", + "Type": "Float", + "Desc": "Minimum sum of instance weight(hessian) needed in a child. If the tree partition step results in a leaf node with the sum of instance weight less than min_child_weight, then the building process will give up further partitioning. In linear regression mode, this simply corresponds to minimum number of instances needed to be in each node. The larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "SubsampleFrequency", + "Type": "Int", + "Desc": "Subsample frequency for bagging. 0 means no subsample. Specifies the frequency at which the bagging occurs, where if this is set to N, the subsampling will happen at every N iterations.This must be set with Subsample as this specifies the amount to subsample.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "SubsampleFraction", + "Type": "Float", + "Desc": "Subsample ratio of the training instance. Setting it to 0.5 means that LightGBM randomly collected half of the data instances to grow trees and this will prevent overfitting. Range: (0,1].", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "Subsample ratio of columns when constructing each tree. Range: (0,1].", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization term on weights, increasing this value will make model more conservative.", + "Aliases": [ + "l2" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization term on weights, increase this value will make model more conservative.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + } + ] + }, + { + "Name": "gbdt", + "Desc": "Traditional Gradient Boosting Decision Tree.", + "FriendlyName": "Tree Booster", + "Settings": [ + { + "Name": "MinimumSplitGain", + "Type": "Float", + "Desc": "Minimum loss reduction required to make a further partition on a leaf node of the tree. the larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "MaximumTreeDepth", + "Type": "Int", + "Desc": "Maximum depth of a tree. 0 means no limit. However, tree still grows by best-first.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "MinimumChildWeight", + "Type": "Float", + "Desc": "Minimum sum of instance weight(hessian) needed in a child. If the tree partition step results in a leaf node with the sum of instance weight less than min_child_weight, then the building process will give up further partitioning. In linear regression mode, this simply corresponds to minimum number of instances needed to be in each node. The larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "SubsampleFrequency", + "Type": "Int", + "Desc": "Subsample frequency for bagging. 0 means no subsample. Specifies the frequency at which the bagging occurs, where if this is set to N, the subsampling will happen at every N iterations.This must be set with Subsample as this specifies the amount to subsample.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "SubsampleFraction", + "Type": "Float", + "Desc": "Subsample ratio of the training instance. Setting it to 0.5 means that LightGBM randomly collected half of the data instances to grow trees and this will prevent overfitting. Range: (0,1].", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "Subsample ratio of columns when constructing each tree. Range: (0,1].", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization term on weights, increasing this value will make model more conservative.", + "Aliases": [ + "l2" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization term on weights, increase this value will make model more conservative.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + } + ] + }, + { + "Name": "goss", + "Desc": "Gradient-based One-Side Sampling.", + "FriendlyName": "Gradient-based One-Size Sampling", + "Settings": [ + { + "Name": "TopRate", + "Type": "Float", + "Desc": "Retain ratio for large gradient instances.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.2, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "OtherRate", + "Type": "Float", + "Desc": "Retain ratio for small gradient instances.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "MinimumSplitGain", + "Type": "Float", + "Desc": "Minimum loss reduction required to make a further partition on a leaf node of the tree. the larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "MaximumTreeDepth", + "Type": "Int", + "Desc": "Maximum depth of a tree. 0 means no limit. However, tree still grows by best-first.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "MinimumChildWeight", + "Type": "Float", + "Desc": "Minimum sum of instance weight(hessian) needed in a child. If the tree partition step results in a leaf node with the sum of instance weight less than min_child_weight, then the building process will give up further partitioning. In linear regression mode, this simply corresponds to minimum number of instances needed to be in each node. The larger, the more conservative the algorithm will be.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.1, + "Range": { + "Min": 0.0 + } + }, + { + "Name": "SubsampleFrequency", + "Type": "Int", + "Desc": "Subsample frequency for bagging. 0 means no subsample. Specifies the frequency at which the bagging occurs, where if this is set to N, the subsampling will happen at every N iterations.This must be set with Subsample as this specifies the amount to subsample.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0, + "Range": { + "Max": 2147483647, + "Min": 0 + } + }, + { + "Name": "SubsampleFraction", + "Type": "Float", + "Desc": "Subsample ratio of the training instance. Setting it to 0.5 means that LightGBM randomly collected half of the data instances to grow trees and this will prevent overfitting. Range: (0,1].", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "Subsample ratio of columns when constructing each tree. Range: (0,1].", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "Range": { + "Inf": 0.0, + "Max": 1.0 + } + }, + { + "Name": "L2Regularization", + "Type": "Float", + "Desc": "L2 regularization term on weights, increasing this value will make model more conservative.", + "Aliases": [ + "l2" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + }, + { + "Name": "L1Regularization", + "Type": "Float", + "Desc": "L1 regularization term on weights, increase this value will make model more conservative.", + "Aliases": [ + "l1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "Range": { + "Min": 0.0 + }, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 0.5, + 1.0 + ] + } + } + ] + } + ] + }, + { + "Kind": "CalibratorTrainer", + "Components": [ + { + "Name": "FixedPlattCalibrator", + "Desc": null, + "FriendlyName": "Fixed Platt Calibrator", + "Aliases": [ + "FixedPlatt", + "FixedSigmoid" + ], + "Settings": [ + { + "Name": "Slope", + "Type": "Float", + "Desc": "The slope parameter of f(x) = 1 / (1 + exp(slope * x + offset)", + "Aliases": [ + "a" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1.0 + }, + { + "Name": "Offset", + "Type": "Float", + "Desc": "The offset parameter of f(x) = 1 / (1 + exp(slope * x + offset)", + "Aliases": [ + "b" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + } + ] + }, + { + "Name": "NaiveCalibrator", + "Desc": null, + "FriendlyName": "Naive Calibrator", + "Aliases": [ + "Naive" + ], + "Settings": [] + }, + { + "Name": "PavCalibrator", + "Desc": null, + "FriendlyName": "PAV Calibrator", + "Aliases": [ + "Pav" + ], + "Settings": [] + }, + { + "Name": "PlattCalibrator", + "Desc": "Platt calibration.", + "FriendlyName": "Platt Calibrator", + "Aliases": [ + "Platt", + "Sigmoid" + ], + "Settings": [] + } + ] + }, + { + "Kind": "ClassificationLossFunction", + "Components": [ + { + "Name": "ExpLoss", + "Desc": "Exponential loss.", + "FriendlyName": "Exponential Loss", + "Settings": [ + { + "Name": "Beta", + "Type": "Float", + "Desc": "Beta (dilation)", + "Aliases": [ + "beta" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ] + }, + { + "Name": "HingeLoss", + "Desc": "Hinge loss.", + "FriendlyName": "Hinge loss", + "Aliases": [ + "Hinge" + ], + "Settings": [ + { + "Name": "Margin", + "Type": "Float", + "Desc": "Margin value", + "Aliases": [ + "marg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ] + }, + { + "Name": "LogLoss", + "Desc": "Log loss.", + "FriendlyName": "Log loss", + "Aliases": [ + "Logistic", + "CrossEntropy" + ], + "Settings": [] + }, + { + "Name": "SmoothedHingeLoss", + "Desc": "Smoothed Hinge loss.", + "FriendlyName": "Smoothed Hinge Loss", + "Aliases": [ + "SmoothedHinge" + ], + "Settings": [ + { + "Name": "SmoothingConst", + "Type": "Float", + "Desc": "Smoothing constant", + "Aliases": [ + "smooth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ] + } + ] + }, + { + "Kind": "CountTableBuilder", + "Components": [ + { + "Name": "CMSketch", + "Desc": "Create the count table using the count-min sketch structure, which has a smaller memory footprint, at the expense of some overcounting due to collisions.", + "FriendlyName": "Count Min Table Builder", + "Aliases": [ + "CMTable" + ], + "Settings": [ + { + "Name": "Depth", + "Type": "Int", + "Desc": "Count-Min Sketch table depth", + "Aliases": [ + "d" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 4 + }, + { + "Name": "Width", + "Type": "Int", + "Desc": "Count-Min Sketch width", + "Aliases": [ + "w" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 8388608 + } + ] + }, + { + "Name": "Dict", + "Desc": "Build a dictionary containing the exact count of each categorical feature value.", + "FriendlyName": "Dictionary Based Count Table Builder", + "Aliases": [ + "Dictionary" + ], + "Settings": [ + { + "Name": "GarbageThreshold", + "Type": "Float", + "Desc": "Garbage threshold (counts below or equal to the threshold are assigned to the garbage bin)", + "Aliases": [ + "gb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + } + ] + } + ] + }, + { + "Kind": "EarlyStoppingCriterion", + "Components": [ + { + "Name": "GL", + "Desc": "Stop in case of loss of generality.", + "FriendlyName": "Loss of Generality (GL)", + "Settings": [ + { + "Name": "Threshold", + "Type": "Float", + "Desc": "Threshold in range [0,1].", + "Aliases": [ + "th" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Max": 1.0, + "Min": 0.0 + } + } + ] + }, + { + "Name": "LP", + "Desc": "Stops in case of low progress.", + "FriendlyName": "Low Progress (LP)", + "Settings": [ + { + "Name": "Threshold", + "Type": "Float", + "Desc": "Threshold in range [0,1].", + "Aliases": [ + "th" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Max": 1.0, + "Min": 0.0 + } + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The window size.", + "Aliases": [ + "w" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5, + "Range": { + "Inf": 0 + } + } + ] + }, + { + "Name": "PQ", + "Desc": "Stops in case of generality to progress ration exceeds threshold.", + "FriendlyName": "Generality to Progress Ratio (PQ)", + "Settings": [ + { + "Name": "Threshold", + "Type": "Float", + "Desc": "Threshold in range [0,1].", + "Aliases": [ + "th" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Max": 1.0, + "Min": 0.0 + } + }, + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The window size.", + "Aliases": [ + "w" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5, + "Range": { + "Inf": 0 + } + } + ] + }, + { + "Name": "TR", + "Desc": "Stop if validation score exceeds threshold value.", + "FriendlyName": "Tolerant (TR)", + "Settings": [ + { + "Name": "Threshold", + "Type": "Float", + "Desc": "Tolerance threshold. (Non negative value)", + "Aliases": [ + "th" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.01, + "Range": { + "Min": 0.0 + } + } + ] + }, + { + "Name": "UP", + "Desc": "Stops in case of consecutive loss in generality.", + "FriendlyName": "Consecutive Loss in Generality (UP)", + "Settings": [ + { + "Name": "WindowSize", + "Type": "Int", + "Desc": "The window size.", + "Aliases": [ + "w" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5, + "Range": { + "Inf": 0 + } + } + ] + } + ] + }, + { + "Kind": "EnsembleBinaryDiversityMeasure", + "Components": [ + { + "Name": "DisagreementDiversityMeasure", + "Desc": null, + "FriendlyName": "Disagreement Diversity Measure", + "Settings": [] + } + ] + }, + { + "Kind": "EnsembleBinaryOutputCombiner", + "Components": [ + { + "Name": "Average", + "Desc": null, + "FriendlyName": "Average", + "Settings": [] + }, + { + "Name": "Median", + "Desc": null, + "FriendlyName": "Median", + "Settings": [] + }, + { + "Name": "Stacking", + "Desc": null, + "FriendlyName": "Stacking", + "Settings": [ + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + }, + { + "Name": "Voting", + "Desc": null, + "FriendlyName": "Voting", + "Settings": [] + }, + { + "Name": "WeightedAverage", + "Desc": null, + "FriendlyName": "Weighted Average", + "Settings": [ + { + "Name": "WeightageName", + "Type": { + "Kind": "Enum", + "Values": [ + "Accuracy", + "Auc", + "PosPrecision", + "PosRecall", + "NegPrecision", + "NegRecall" + ] + }, + "Desc": "The metric type to be used to find the weights for each model", + "Aliases": [ + "wn" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": "Auc" + } + ] + } + ] + }, + { + "Kind": "EnsembleBinarySubModelSelector", + "Components": [ + { + "Name": "AllSelector", + "Desc": null, + "FriendlyName": "All Selector", + "Settings": [] + }, + { + "Name": "BestDiverseSelector", + "Desc": null, + "FriendlyName": "Best Diverse Selector", + "Settings": [ + { + "Name": "DiversityMetricType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleBinaryDiversityMeasure" + }, + "Desc": "The metric type to be used to find the diversity among base learners", + "Aliases": [ + "dm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "DisagreementDiversityMeasure" + } + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + }, + { + "Name": "BestPerformanceSelector", + "Desc": null, + "FriendlyName": "Best Performance Selector", + "Settings": [ + { + "Name": "MetricName", + "Type": { + "Kind": "Enum", + "Values": [ + "Accuracy", + "PosPrecName", + "PosRecallName", + "NegPrecName", + "NegRecallName", + "Auc", + "LogLoss", + "LogLossReduction", + "F1", + "AuPrc" + ] + }, + "Desc": "The metric type to be used to find the best performance", + "Aliases": [ + "mn" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": "Auc" + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + } + ] + }, + { + "Kind": "EnsembleFeatureSelector", + "Components": [ + { + "Name": "AllFeatureSelector", + "Desc": null, + "FriendlyName": "All Feature Selector", + "Settings": [] + }, + { + "Name": "RandomFeatureSelector", + "Desc": null, + "FriendlyName": "Random Feature Selector", + "Settings": [ + { + "Name": "FeaturesSelectionProportion", + "Type": "Float", + "Desc": "The proportion of features to be selected. The range is 0.0-1.0", + "Aliases": [ + "fp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.8 + } + ] + } + ] + }, + { + "Kind": "EnsembleMulticlassDiversityMeasure", + "Components": [ + { + "Name": "MultiDisagreementDiversityMeasure", + "Desc": null, + "FriendlyName": "Disagreement Diversity Measure", + "Settings": [] + } + ] + }, + { + "Kind": "EnsembleMulticlassOutputCombiner", + "Components": [ + { + "Name": "MultiAverage", + "Desc": null, + "FriendlyName": "Average", + "Settings": [ + { + "Name": "Normalize", + "Type": "Bool", + "Desc": "Whether to normalize the output of base models before combining them", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": true + } + ] + }, + { + "Name": "MultiMedian", + "Desc": null, + "FriendlyName": "Median", + "Settings": [ + { + "Name": "Normalize", + "Type": "Bool", + "Desc": "Whether to normalize the output of base models before combining them", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": true + } + ] + }, + { + "Name": "MultiStacking", + "Desc": null, + "FriendlyName": "Stacking", + "Settings": [ + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + }, + { + "Name": "MultiVoting", + "Desc": null, + "FriendlyName": "Voting", + "Settings": [] + }, + { + "Name": "MultiWeightedAverage", + "Desc": null, + "FriendlyName": "Multi Weighted Average", + "Settings": [ + { + "Name": "WeightageName", + "Type": { + "Kind": "Enum", + "Values": [ + "AccuracyMicroAvg", + "AccuracyMacroAvg" + ] + }, + "Desc": "The metric type to be used to find the weights for each model", + "Aliases": [ + "wn" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": "AccuracyMicroAvg" + }, + { + "Name": "Normalize", + "Type": "Bool", + "Desc": "Whether to normalize the output of base models before combining them", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": true + } + ] + } + ] + }, + { + "Kind": "EnsembleMulticlassSubModelSelector", + "Components": [ + { + "Name": "AllSelectorMultiClass", + "Desc": null, + "FriendlyName": "All Selector", + "Settings": [] + }, + { + "Name": "BestDiverseSelectorMultiClass", + "Desc": null, + "FriendlyName": "Best Diverse Selector", + "Settings": [ + { + "Name": "DiversityMetricType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleMulticlassDiversityMeasure" + }, + "Desc": "The metric type to be used to find the diversity among base learners", + "Aliases": [ + "dm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "MultiDisagreementDiversityMeasure" + } + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + }, + { + "Name": "BestPerformanceSelectorMultiClass", + "Desc": null, + "FriendlyName": "Best Performance Selector", + "Settings": [ + { + "Name": "MetricName", + "Type": { + "Kind": "Enum", + "Values": [ + "AccuracyMicro", + "AccuracyMacro", + "LogLoss", + "LogLossReduction" + ] + }, + "Desc": "The metric type to be used to find the best performance", + "Aliases": [ + "mn" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": "AccuracyMicro" + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + } + ] + }, + { + "Kind": "EnsembleRegressionDiversityMeasure", + "Components": [ + { + "Name": "RegressionDisagreementDiversityMeasure", + "Desc": null, + "FriendlyName": "Disagreement Diversity Measure", + "Settings": [] + } + ] + }, + { + "Kind": "EnsembleRegressionOutputCombiner", + "Components": [ + { + "Name": "Average", + "Desc": null, + "FriendlyName": "Average", + "Settings": [] + }, + { + "Name": "Median", + "Desc": null, + "FriendlyName": "Median", + "Settings": [] + }, + { + "Name": "RegressionStacking", + "Desc": null, + "FriendlyName": "Stacking", + "Settings": [ + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + } + ] + }, + { + "Kind": "EnsembleRegressionSubModelSelector", + "Components": [ + { + "Name": "AllSelector", + "Desc": null, + "FriendlyName": "All Selector", + "Settings": [] + }, + { + "Name": "BestDiverseSelectorRegression", + "Desc": null, + "FriendlyName": "Best Diverse Selector", + "Settings": [ + { + "Name": "DiversityMetricType", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleRegressionDiversityMeasure" + }, + "Desc": "The metric type to be used to find the diversity among base learners", + "Aliases": [ + "dm" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": { + "Name": "RegressionDisagreementDiversityMeasure" + } + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + }, + { + "Name": "BestPerformanceRegressionSelector", + "Desc": null, + "FriendlyName": "Best Performance Selector", + "Settings": [ + { + "Name": "MetricName", + "Type": { + "Kind": "Enum", + "Values": [ + "L1", + "L2", + "Rms", + "Loss", + "RSquared" + ] + }, + "Desc": "The metric type to be used to find the best performance", + "Aliases": [ + "mn" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": "L1" + }, + { + "Name": "LearnersSelectionProportion", + "Type": "Float", + "Desc": "The proportion of best base learners to be selected. The range is 0.0-1.0", + "Aliases": [ + "lp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.5 + }, + { + "Name": "ValidationDatasetProportion", + "Type": "Float", + "Desc": "The proportion of instances to be selected to test the individual base learner. If it is 0, it uses training set", + "Aliases": [ + "vp" + ], + "Required": false, + "SortOrder": 50.0, + "IsNullable": false, + "Default": 0.3 + } + ] + } + ] + }, + { + "Kind": "EnsembleSubsetSelector", + "Components": [ + { + "Name": "AllInstanceSelector", + "Desc": null, + "FriendlyName": "All Instance Selector", + "Settings": [ + { + "Name": "FeatureSelector", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleFeatureSelector" + }, + "Desc": "The Feature selector", + "Aliases": [ + "fs" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": { + "Name": "AllFeatureSelector" + } + } + ] + }, + { + "Name": "BootstrapSelector", + "Desc": null, + "FriendlyName": "Bootstrap Selector", + "Settings": [ + { + "Name": "FeatureSelector", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleFeatureSelector" + }, + "Desc": "The Feature selector", + "Aliases": [ + "fs" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": { + "Name": "AllFeatureSelector" + } + } + ] + }, + { + "Name": "RandomPartitionSelector", + "Desc": null, + "FriendlyName": "Random Partition Selector", + "Settings": [ + { + "Name": "FeatureSelector", + "Type": { + "Kind": "Component", + "ComponentKind": "EnsembleFeatureSelector" + }, + "Desc": "The Feature selector", + "Aliases": [ + "fs" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": { + "Name": "AllFeatureSelector" + } + } + ] + } + ] + }, + { + "Kind": "FastTreeTrainer", + "Components": [ + { + "Name": "FastTreeBinaryClassification", + "Desc": "Uses a logit-boost boosted tree learner to perform binary classification.", + "FriendlyName": "FastTree (Boosted Trees) Classification", + "Settings": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "UnbalancedSets", + "Type": "Bool", + "Desc": "Option for using derivatives optimized for unbalanced sets", + "Aliases": [ + "us" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ] + }, + { + "Name": "FastTreeRanking", + "Desc": "Trains gradient boosted decision trees to the LambdaRank quasi-gradient.", + "FriendlyName": "FastTree (Boosted Trees) Ranking", + "Settings": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "GroupId" + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "CustomGains", + "Type": { + "Kind": "Array", + "ItemType": "Float" + }, + "Desc": "Comma-separated list of gains associated to each relevance label.", + "Aliases": [ + "gains" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + 0.0, + 3.0, + 7.0, + 15.0, + 31.0 + ] + }, + { + "Name": "UseDcg", + "Type": "Bool", + "Desc": "Train DCG instead of NDCG", + "Aliases": [ + "dcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "SortingAlgorithm", + "Type": "String", + "Desc": "The sorting algorithm to use for DCG and LambdaMart calculations [DescendingStablePessimistic/DescendingStable/DescendingReverse/DescendingDotNet]", + "Aliases": [ + "sort" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "DescendingStablePessimistic" + }, + { + "Name": "NdcgTruncationLevel", + "Type": "Int", + "Desc": "max-NDCG truncation to use in the LambdaMART algorithm", + "Aliases": [ + "n" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "ShiftedNdcg", + "Type": "Bool", + "Desc": "Use shifted NDCG", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "CostFunctionParam", + "Type": "Char", + "Desc": "Cost function parameter (w/c)", + "Aliases": [ + "cf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "w" + }, + { + "Name": "DistanceWeight2", + "Type": "Bool", + "Desc": "Distance weight 2 adjustment to cost", + "Aliases": [ + "dw" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "NormalizeQueryLambdas", + "Type": "Bool", + "Desc": "Normalize query lambdas", + "Aliases": [ + "nql" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ] + }, + { + "Name": "FastTreeRegression", + "Desc": "Trains gradient boosted decision trees to fit target values using least-squares.", + "FriendlyName": "FastTree (Boosted Trees) Regression", + "Settings": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ] + }, + { + "Name": "FastTreeTweedieRegression", + "Desc": "Trains gradient boosted decision trees to fit target values using a Tweedie loss function. This learner is a generalization of Poisson, compound Poisson, and gamma regression.", + "FriendlyName": "FastTree (Boosted Trees) Tweedie Regression", + "Settings": [ + { + "Name": "NumberOfTrees", + "Type": "Int", + "Desc": "Total number of decision trees to create in the ensemble", + "Aliases": [ + "iter" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": 100, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 20, + 100, + 500 + ] + } + }, + { + "Name": "TrainingData", + "Type": "DataView", + "Desc": "The data to be used for training", + "Aliases": [ + "data" + ], + "Required": true, + "SortOrder": 1.0, + "IsNullable": false + }, + { + "Name": "NumberOfLeaves", + "Type": "Int", + "Desc": "The max number of leaves in each regression tree", + "Aliases": [ + "nl" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 20, + "SweepRange": { + "RangeType": "Long", + "Min": 2, + "Max": 128, + "StepSize": 4.0, + "IsLogScale": true + } + }, + { + "Name": "FeatureColumnName", + "Type": "String", + "Desc": "Column to use for features", + "Aliases": [ + "feat" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": "Features" + }, + { + "Name": "MinimumExampleCountPerLeaf", + "Type": "Int", + "Desc": "The minimal number of examples allowed in a leaf of a regression tree, out of the subsampled data", + "Aliases": [ + "mil" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 10, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 1, + 10, + 50 + ] + } + }, + { + "Name": "LabelColumnName", + "Type": "String", + "Desc": "Column to use for labels", + "Aliases": [ + "lab" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": "Label" + }, + { + "Name": "LearningRate", + "Type": "Float", + "Desc": "The learning rate", + "Aliases": [ + "lr" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0.2, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 0.4, + "IsLogScale": true + } + }, + { + "Name": "ExampleWeightColumnName", + "Type": "String", + "Desc": "Column to use for example weight", + "Aliases": [ + "weight" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "RowGroupColumnName", + "Type": "String", + "Desc": "Column to use for example groupId", + "Aliases": [ + "groupId" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "NormalizeFeatures", + "Type": { + "Kind": "Enum", + "Values": [ + "No", + "Warn", + "Auto", + "Yes" + ] + }, + "Desc": "Normalize option for the feature column", + "Aliases": [ + "norm" + ], + "Required": false, + "SortOrder": 5.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Caching", + "Type": { + "Kind": "Enum", + "Values": [ + "Auto", + "Memory", + "None" + ] + }, + "Desc": "Whether trainer should cache input training data", + "Aliases": [ + "cache" + ], + "Required": false, + "SortOrder": 6.0, + "IsNullable": false, + "Default": "Auto" + }, + { + "Name": "Index", + "Type": "Float", + "Desc": "Index parameter for the Tweedie distribution, in the range [1, 2]. 1 is Poisson loss, 2 is gamma loss, and intermediate values are compound Poisson loss.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.5 + }, + { + "Name": "BestStepRankingRegressionTrees", + "Type": "Bool", + "Desc": "Option for using best regression step trees", + "Aliases": [ + "bsr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseLineSearch", + "Type": "Bool", + "Desc": "Should we use line search for a step size", + "Aliases": [ + "ls" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumNumberOfLineSearchSteps", + "Type": "Int", + "Desc": "Number of post-bracket line search steps", + "Aliases": [ + "lssteps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "MinimumStepSize", + "Type": "Float", + "Desc": "Minimum line search step size", + "Aliases": [ + "minstep" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "OptimizationAlgorithm", + "Type": { + "Kind": "Enum", + "Values": [ + "GradientDescent", + "AcceleratedGradientDescent", + "ConjugateGradientDescent" + ] + }, + "Desc": "Optimization algorithm to be used (GradientDescent, AcceleratedGradientDescent)", + "Aliases": [ + "oa" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "GradientDescent" + }, + { + "Name": "EarlyStoppingRule", + "Type": { + "Kind": "Component", + "ComponentKind": "EarlyStoppingCriterion" + }, + "Desc": "Early stopping rule. (Validation set (/valid) is required.)", + "Aliases": [ + "esr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "EarlyStoppingMetrics", + "Type": "Int", + "Desc": "Early stopping metrics. (For regression, 1: L1, 2:L2; for ranking, 1:NDCG@1, 3:NDCG@3)", + "Aliases": [ + "esmt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "EnablePruning", + "Type": "Bool", + "Desc": "Enable post-training pruning to avoid overfitting. (a validation set is required)", + "Aliases": [ + "pruning" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "UseTolerantPruning", + "Type": "Bool", + "Desc": "Use window and tolerance for pruning", + "Aliases": [ + "prtol" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PruningThreshold", + "Type": "Float", + "Desc": "The tolerance threshold for pruning", + "Aliases": [ + "prth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.004 + }, + { + "Name": "PruningWindowSize", + "Type": "Int", + "Desc": "The moving window size for pruning", + "Aliases": [ + "prws" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 5 + }, + { + "Name": "Shrinkage", + "Type": "Float", + "Desc": "Shrinkage", + "Aliases": [ + "shrk" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0, + "SweepRange": { + "RangeType": "Float", + "Min": 0.025, + "Max": 4.0, + "IsLogScale": true + } + }, + { + "Name": "DropoutRate", + "Type": "Float", + "Desc": "Dropout rate for tree regularization", + "Aliases": [ + "tdrop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0, + "SweepRange": { + "RangeType": "Discrete", + "Values": [ + 0.0, + 1E-09, + 0.05, + 0.1, + 0.2 + ] + } + }, + { + "Name": "GetDerivativesSampleRate", + "Type": "Int", + "Desc": "Sample each query 1 in k times in the GetDerivatives function", + "Aliases": [ + "sr" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "WriteLastEnsemble", + "Type": "Bool", + "Desc": "Write the last ensemble instead of the one determined by early stopping", + "Aliases": [ + "hl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumTreeOutput", + "Type": "Float", + "Desc": "Upper bound on absolute value of single tree output", + "Aliases": [ + "mo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100.0 + }, + { + "Name": "RandomStart", + "Type": "Bool", + "Desc": "Training starts from random ordering (determined by /r1)", + "Aliases": [ + "rs" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "FilterZeroLambdas", + "Type": "Bool", + "Desc": "Filter zero lambdas during training", + "Aliases": [ + "fzl" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "BaselineScoresFormula", + "Type": "String", + "Desc": "Freeform defining the scores that should be used as the baseline ranker", + "Aliases": [ + "basescores" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "BaselineAlphaRisk", + "Type": "String", + "Desc": "Baseline alpha for tradeoffs of risk (0 is normal training)", + "Aliases": [ + "basealpha" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "PositionDiscountFreeform", + "Type": "String", + "Desc": "The discount freeform which specifies the per position discounts of examples in a query (uses a single variable P for position where P=0 is first position)", + "Aliases": [ + "pdff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "ParallelTrainer", + "Type": { + "Kind": "Component", + "ComponentKind": "ParallelTraining" + }, + "Desc": "Allows to choose Parallel FastTree Learning Algorithm", + "Aliases": [ + "parag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": { + "Name": "Single" + } + }, + { + "Name": "NumberOfThreads", + "Type": "Int", + "Desc": "The number of threads to use", + "Aliases": [ + "t" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Seed", + "Type": "Int", + "Desc": "The seed of the random number generator", + "Aliases": [ + "r1" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "FeatureSelectionSeed", + "Type": "Int", + "Desc": "The seed of the active feature selection", + "Aliases": [ + "r3" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 123 + }, + { + "Name": "EntropyCoefficient", + "Type": "Float", + "Desc": "The entropy (regularization) coefficient between 0 and 1", + "Aliases": [ + "e" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "HistogramPoolSize", + "Type": "Int", + "Desc": "The number of histograms in the pool (between 2 and numLeaves)", + "Aliases": [ + "ps" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": -1 + }, + { + "Name": "DiskTranspose", + "Type": "Bool", + "Desc": "Whether to utilize the disk or the data's native transposition facilities (where applicable) when performing the transpose", + "Aliases": [ + "dt" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "FeatureFlocks", + "Type": "Bool", + "Desc": "Whether to collectivize features during dataset preparation to speed up training", + "Aliases": [ + "flocks" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "CategoricalSplit", + "Type": "Bool", + "Desc": "Whether to do split based on multiple categorical feature values.", + "Aliases": [ + "cat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MaximumCategoricalGroupCountPerNode", + "Type": "Int", + "Desc": "Maximum categorical split groups to consider when splitting on a categorical feature. Split groups are a collection of split points. This is used to reduce overfitting when there many categorical features.", + "Aliases": [ + "mcg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MaximumCategoricalSplitPointCount", + "Type": "Int", + "Desc": "Maximum categorical split points to consider when splitting on a categorical feature.", + "Aliases": [ + "maxcat" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 64 + }, + { + "Name": "MinimumExampleFractionForCategoricalSplit", + "Type": "Float", + "Desc": "Minimum categorical example percentage in a bin to consider for a split.", + "Aliases": [ + "mdop" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.001 + }, + { + "Name": "MinimumExamplesForCategoricalSplit", + "Type": "Int", + "Desc": "Minimum categorical example count in a bin to consider for a split.", + "Aliases": [ + "mdo" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 100 + }, + { + "Name": "Bias", + "Type": "Float", + "Desc": "Bias for calculating gradient for each feature bin for a categorical feature.", + "Aliases": [ + "bias" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "Bundling", + "Type": { + "Kind": "Enum", + "Values": [ + "None", + "AggregateLowPopulation", + "Adjacent" + ] + }, + "Desc": "Bundle low population bins. Bundle.None(0): no bundling, Bundle.AggregateLowPopulation(1): Bundle low population, Bundle.Adjacent(2): Neighbor low population bundle.", + "Aliases": [ + "bundle" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "None" + }, + { + "Name": "MaximumBinCountPerFeature", + "Type": "Int", + "Desc": "Maximum number of distinct values (bins) per feature", + "Aliases": [ + "mb" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 255 + }, + { + "Name": "SparsifyThreshold", + "Type": "Float", + "Desc": "Sparsity level needed to use sparse feature representation", + "Aliases": [ + "sp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFirstUsePenalty", + "Type": "Float", + "Desc": "The feature first use penalty coefficient", + "Aliases": [ + "ffup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "FeatureReusePenalty", + "Type": "Float", + "Desc": "The feature re-use penalty (regularization) coefficient", + "Aliases": [ + "frup" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "GainConfidenceLevel", + "Type": "Float", + "Desc": "Tree fitting gain confidence requirement (should be in the range [0,1) ).", + "Aliases": [ + "gainconf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "SoftmaxTemperature", + "Type": "Float", + "Desc": "The temperature of the randomized softmax distribution for choosing the feature", + "Aliases": [ + "smtemp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "ExecutionTime", + "Type": "Bool", + "Desc": "Print execution time breakdown to stdout", + "Aliases": [ + "et" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "MemoryStatistics", + "Type": "Bool", + "Desc": "Print memory statistics to stdout", + "Aliases": [ + "memstats" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureFraction", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each iteration", + "Aliases": [ + "ff" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "BaggingSize", + "Type": "Int", + "Desc": "Number of trees in each bag (0 for disabling bagging)", + "Aliases": [ + "bag" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "BaggingExampleFraction", + "Type": "Float", + "Desc": "Percentage of training examples used in each bag", + "Aliases": [ + "bagfrac" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.7 + }, + { + "Name": "FeatureFractionPerSplit", + "Type": "Float", + "Desc": "The fraction of features (chosen randomly) to use on each split", + "Aliases": [ + "sf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + }, + { + "Name": "Smoothing", + "Type": "Float", + "Desc": "Smoothing paramter for tree regularization", + "Aliases": [ + "s" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0.0 + }, + { + "Name": "AllowEmptyTrees", + "Type": "Bool", + "Desc": "When a root split is impossible, allow training to proceed", + "Aliases": [ + "allowempty", + "dummies" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "FeatureCompressionLevel", + "Type": "Int", + "Desc": "The level of feature compression to use", + "Aliases": [ + "fcomp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "CompressEnsemble", + "Type": "Bool", + "Desc": "Compress the tree Ensemble", + "Aliases": [ + "cmp" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTestGraph", + "Type": "Bool", + "Desc": "Print metrics graph for the first test set", + "Aliases": [ + "graph" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "PrintTrainValidGraph", + "Type": "Bool", + "Desc": "Print Train and Validation metrics in graph", + "Aliases": [ + "graphtv" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": false + }, + { + "Name": "TestFrequency", + "Type": "Int", + "Desc": "Calculate metric values for train/valid/test every k rounds", + "Aliases": [ + "tf" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 2147483647 + } + ] + } + ] + }, + { + "Kind": "NgramExtractor", + "Components": [ + { + "Name": "NGram", + "Desc": "Extracts NGrams from text and convert them to vector using dictionary.", + "FriendlyName": "NGram Extractor Transform", + "Aliases": [ + "NGramExtractorTransform", + "NGramExtractor" + ], + "Settings": [ + { + "Name": "NgramLength", + "Type": "Int", + "Desc": "Ngram length", + "Aliases": [ + "ngram" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "SkipLength", + "Type": "Int", + "Desc": "Maximum number of tokens to skip when constructing an n-gram", + "Aliases": [ + "skips" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "AllLengths", + "Type": "Bool", + "Desc": "Whether to include all n-gram lengths up to NgramLength or only NgramLength", + "Aliases": [ + "all" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaxNumTerms", + "Type": { + "Kind": "Array", + "ItemType": "Int" + }, + "Desc": "Maximum number of n-grams to store in the dictionary", + "Aliases": [ + "max" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": [ + 10000000 + ] + }, + { + "Name": "Weighting", + "Type": { + "Kind": "Enum", + "Values": [ + "Tf", + "Idf", + "TfIdf" + ] + }, + "Desc": "The weighting criteria", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "Tf" + }, + { + "Name": "TermSeparator", + "Type": "Char", + "Desc": "Separator used to separate terms/frequency pairs.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "\u0000" + }, + { + "Name": "FreqSeparator", + "Type": "Char", + "Desc": "Separator used to separate terms from their frequency.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "\u0000" + } + ] + }, + { + "Name": "NGramHash", + "Desc": "Extracts NGrams from text and convert them to vector using hashing trick.", + "FriendlyName": "NGram Hash Extractor Transform", + "Aliases": [ + "NGramHashExtractorTransform", + "NGramHashExtractor" + ], + "Settings": [ + { + "Name": "NumberOfBits", + "Type": "Int", + "Desc": "Number of bits to hash into. Must be between 1 and 30, inclusive.", + "Aliases": [ + "bits" + ], + "Required": false, + "SortOrder": 2.0, + "IsNullable": false, + "Default": 16 + }, + { + "Name": "NgramLength", + "Type": "Int", + "Desc": "Ngram length", + "Aliases": [ + "ngram" + ], + "Required": false, + "SortOrder": 3.0, + "IsNullable": false, + "Default": 1 + }, + { + "Name": "SkipLength", + "Type": "Int", + "Desc": "Maximum number of tokens to skip when constructing an n-gram", + "Aliases": [ + "skips" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": 0 + }, + { + "Name": "AllLengths", + "Type": "Bool", + "Desc": "Whether to include all n-gram lengths up to ngramLength or only ngramLength", + "Aliases": [ + "all" + ], + "Required": false, + "SortOrder": 4.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "Seed", + "Type": "UInt", + "Desc": "Hashing seed", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 314489979 + }, + { + "Name": "Ordered", + "Type": "Bool", + "Desc": "Whether the position of each source column should be included in the hash (when there are multiple source columns).", + "Aliases": [ + "ord" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": true + }, + { + "Name": "MaximumNumberOfInverts", + "Type": "Int", + "Desc": "Limit the number of keys used to generate the slot name to this many. 0 means no invert hashing, -1 means no limit.", + "Aliases": [ + "ih" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ] + } + ] + }, + { + "Kind": "ParallelLightGBM", + "Components": [ + { + "Name": "Single", + "Desc": "Single node machine learning process.", + "FriendlyName": "Single", + "Settings": [] + } + ] + }, + { + "Kind": "ParallelTraining", + "Components": [ + { + "Name": "Single", + "Desc": "Single node machine learning process.", + "FriendlyName": "Single", + "Settings": [] + } + ] + }, + { + "Kind": "PartitionedPathParser", + "Components": [ + { + "Name": "ParquetPathParser", + "Desc": "Extract name/value pairs from Parquet formatted directory names. Example path: Year=2018/Month=12/data1.parquet", + "FriendlyName": "Parquet Partitioned Path Parser", + "Aliases": [ + "ParqPP" + ], + "Settings": [] + }, + { + "Name": "SimplePathParser", + "Desc": "A simple parser that extracts directory names as column values. Column names are defined as arguments.", + "FriendlyName": "Simple Partitioned Path Parser", + "Aliases": [ + "SmplPP" + ], + "Settings": [ + { + "Name": "Columns", + "Type": { + "Kind": "Array", + "ItemType": { + "Kind": "Struct", + "Fields": [ + { + "Name": "Name", + "Type": "String", + "Desc": "Name of the column.", + "Required": true, + "SortOrder": 150.0, + "IsNullable": false + }, + { + "Name": "Type", + "Type": { + "Kind": "Enum", + "Values": [ + "I1", + "U1", + "I2", + "U2", + "I4", + "U4", + "I8", + "U8", + "R4", + "Num", + "R8", + "TX", + "Text", + "TXT", + "BL", + "Bool", + "TimeSpan", + "TS", + "DT", + "DateTime", + "DZ", + "DateTimeZone", + "UG", + "U16" + ] + }, + "Desc": "Data type of the column.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": true, + "Default": null + }, + { + "Name": "Source", + "Type": "Int", + "Desc": "Index of the directory representing this column.", + "Required": true, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 0 + } + ] + } + }, + "Desc": "Column definitions used to override the Partitioned Path Parser. Expected with the format name:type:numeric-source, for example, col=MyFeature:R4:1", + "Aliases": [ + "col" + ], + "Required": false, + "SortOrder": 1.0, + "IsNullable": false, + "Default": null + }, + { + "Name": "Type", + "Type": { + "Kind": "Enum", + "Values": [ + "I1", + "U1", + "I2", + "U2", + "I4", + "U4", + "I8", + "U8", + "R4", + "Num", + "R8", + "TX", + "Text", + "TXT", + "BL", + "Bool", + "TimeSpan", + "TS", + "DT", + "DateTime", + "DZ", + "DateTimeZone", + "UG", + "U16" + ] + }, + "Desc": "Data type of each column.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": "TX" + } + ] + } + ] + }, + { + "Kind": "RegressionLossFunction", + "Components": [ + { + "Name": "PoissonLoss", + "Desc": "Poisson loss.", + "FriendlyName": "Poisson Loss", + "Settings": [] + }, + { + "Name": "SquaredLoss", + "Desc": "Squared loss.", + "FriendlyName": "Squared Loss", + "Aliases": [ + "L2" + ], + "Settings": [] + }, + { + "Name": "TweedieLoss", + "Desc": "Tweedie loss.", + "FriendlyName": "Tweedie Loss", + "Aliases": [ + "tweedie" + ], + "Settings": [ + { + "Name": "Index", + "Type": "Float", + "Desc": "Index parameter for the Tweedie distribution, in the range [1, 2]. 1 is Poisson loss, 2 is gamma loss, and intermediate values are compound Poisson loss.", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.5 + } + ] + } + ] + }, + { + "Kind": "SDCAClassificationLossFunction", + "Components": [ + { + "Name": "HingeLoss", + "Desc": "Hinge loss.", + "FriendlyName": "Hinge loss", + "Aliases": [ + "Hinge" + ], + "Settings": [ + { + "Name": "Margin", + "Type": "Float", + "Desc": "Margin value", + "Aliases": [ + "marg" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ] + }, + { + "Name": "LogLoss", + "Desc": "Log loss.", + "FriendlyName": "Log loss", + "Aliases": [ + "Logistic", + "CrossEntropy" + ], + "Settings": [] + }, + { + "Name": "SmoothedHingeLoss", + "Desc": "Smoothed Hinge loss.", + "FriendlyName": "Smoothed Hinge Loss", + "Aliases": [ + "SmoothedHinge" + ], + "Settings": [ + { + "Name": "SmoothingConst", + "Type": "Float", + "Desc": "Smoothing constant", + "Aliases": [ + "smooth" + ], + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": 1.0 + } + ] + } + ] + }, + { + "Kind": "SDCARegressionLossFunction", + "Components": [ + { + "Name": "SquaredLoss", + "Desc": "Squared loss.", + "FriendlyName": "Squared Loss", + "Aliases": [ + "L2" + ], + "Settings": [] + } + ] + }, + { + "Kind": "StopWordsRemover", + "Components": [ + { + "Name": "Custom", + "Desc": "Remover with list of stopwords specified by the user.", + "FriendlyName": "Custom Stopwords Remover", + "Aliases": [ + "CustomStopWordsRemover", + "CustomStopWords" + ], + "Settings": [ + { + "Name": "Stopword", + "Type": { + "Kind": "Array", + "ItemType": "String" + }, + "Desc": "List of stopwords", + "Required": false, + "SortOrder": 150.0, + "IsNullable": false, + "Default": null + } + ] + }, + { + "Name": "Predefined", + "Desc": "Remover with predefined list of stop words.", + "FriendlyName": "Predefined Stopwords List Remover", + "Aliases": [ + "PredefinedStopWordsRemover", + "PredefinedStopWords" + ], + "Settings": [] + } + ] + } + ], + "EntryPointKinds": [ + { + "Kind": "IAnomalyDetectionOutput", + "Settings": [] + }, + { + "Kind": "IBinaryClassificationOutput", + "Settings": [] + }, + { + "Kind": "ICalibratorInput", + "Settings": [ + { + "Name": "UncalibratedPredictorModel", + "Type": "PredictorModel" + }, + { + "Name": "MaxRows", + "Type": "Int" + }, + { + "Name": "Data", + "Type": "DataView" + } + ] + }, + { + "Kind": "ICalibratorOutput", + "Settings": [] + }, + { + "Kind": "IClassificationEvaluatorOutput", + "Settings": [ + { + "Name": "ConfusionMatrix", + "Type": "DataView" + }, + { + "Name": "Warnings", + "Type": "DataView" + }, + { + "Name": "OverallMetrics", + "Type": "DataView" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView" + } + ] + }, + { + "Kind": "IClusteringOutput", + "Settings": [] + }, + { + "Kind": "IEvaluatorInput", + "Settings": [ + { + "Name": "Data", + "Type": "DataView" + }, + { + "Name": "NameColumn", + "Type": "String" + } + ] + }, + { + "Kind": "IEvaluatorOutput", + "Settings": [ + { + "Name": "Warnings", + "Type": "DataView" + }, + { + "Name": "OverallMetrics", + "Type": "DataView" + }, + { + "Name": "PerInstanceMetrics", + "Type": "DataView" + } + ] + }, + { + "Kind": "IFeaturizerInput", + "Settings": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel" + }, + { + "Name": "Data", + "Type": "DataView" + } + ] + }, + { + "Kind": "IMulticlassClassificationOutput", + "Settings": [] + }, + { + "Kind": "IRankingOutput", + "Settings": [] + }, + { + "Kind": "IRegressionOutput", + "Settings": [] + }, + { + "Kind": "ITrainerInput", + "Settings": [ + { + "Name": "TrainingData", + "Type": "DataView" + }, + { + "Name": "FeatureColumn", + "Type": "String" + } + ] + }, + { + "Kind": "ITrainerInputWithGroupId", + "Settings": [ + { + "Name": "GroupIdColumn", + "Type": "String" + }, + { + "Name": "WeightColumn", + "Type": "String" + }, + { + "Name": "LabelColumn", + "Type": "String" + }, + { + "Name": "TrainingData", + "Type": "DataView" + }, + { + "Name": "FeatureColumn", + "Type": "String" + } + ] + }, + { + "Kind": "ITrainerInputWithLabel", + "Settings": [ + { + "Name": "LabelColumn", + "Type": "String" + }, + { + "Name": "TrainingData", + "Type": "DataView" + }, + { + "Name": "FeatureColumn", + "Type": "String" + } + ] + }, + { + "Kind": "ITrainerInputWithWeight", + "Settings": [ + { + "Name": "WeightColumn", + "Type": "String" + }, + { + "Name": "LabelColumn", + "Type": "String" + }, + { + "Name": "TrainingData", + "Type": "DataView" + }, + { + "Name": "FeatureColumn", + "Type": "String" + } + ] + }, + { + "Kind": "ITrainerOutput", + "Settings": [ + { + "Name": "PredictorModel", + "Type": "PredictorModel" + } + ] + }, + { + "Kind": "ITransformInput", + "Settings": [ + { + "Name": "Data", + "Type": "DataView" + } + ] + }, + { + "Kind": "ITransformOutput", + "Settings": [ + { + "Name": "OutputData", + "Type": "DataView" + }, + { + "Name": "Model", + "Type": "TransformModel" + } + ] + }, + { + "Kind": "IUnsupervisedTrainerWithWeight", + "Settings": [ + { + "Name": "WeightColumn", + "Type": "String" + }, + { + "Name": "TrainingData", + "Type": "DataView" + }, + { + "Name": "FeatureColumn", + "Type": "String" + } + ] + } + ] +} \ No newline at end of file From 67e3150c9a350f60afa7446cfd7a2c31056a59dd Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Thu, 28 Mar 2024 19:58:30 -0700 Subject: [PATCH 22/26] Use test script for helix payload Test script saves test output on failures. It also avoids returning non-zero when the runner completes gracefully with failures. --- eng/Version.Details.xml | 4 ++ eng/Versions.props | 1 + eng/testing/helix.targets | 100 +++++++++++++++++++++++++++++++++++ eng/testing/runTests.cmd | 42 +++++++++++++++ eng/testing/runTests.sh | 38 +++++++++++++ test/Directory.Build.targets | 85 +---------------------------- 6 files changed, 186 insertions(+), 84 deletions(-) create mode 100644 eng/testing/helix.targets create mode 100644 eng/testing/runTests.cmd create mode 100644 eng/testing/runTests.sh diff --git a/eng/Version.Details.xml b/eng/Version.Details.xml index b384be05ab..f97cdebbf2 100644 --- a/eng/Version.Details.xml +++ b/eng/Version.Details.xml @@ -15,6 +15,10 @@ https://github.com/dotnet/arcade 812d978c303174dc1aa305d7359e79053d7d4971 + + https://github.com/dotnet/arcade + 812d978c303174dc1aa305d7359e79053d7d4971 + https://github.com/dotnet/arcade 812d978c303174dc1aa305d7359e79053d7d4971 diff --git a/eng/Versions.props b/eng/Versions.props index 27f26cf54a..d7bc0adb61 100644 --- a/eng/Versions.props +++ b/eng/Versions.props @@ -78,6 +78,7 @@ 8.0.1 5.10.2 1.1.2-beta1.23431.1 + 9.0.0-beta.24165.3 9.0.0-beta.24165.3 9.0.0-beta.24165.3 2.1.0 diff --git a/eng/testing/helix.targets b/eng/testing/helix.targets new file mode 100644 index 0000000000..9cd387278d --- /dev/null +++ b/eng/testing/helix.targets @@ -0,0 +1,100 @@ + + + + + true + + true + + $HELIX_CORRELATION_PAYLOAD + %HELIX_CORRELATION_PAYLOAD% + + runTests.sh + runTests.cmd + + dotnet exec --roll-forward Major --runtimeconfig $(TargetName).runtimeconfig.json --depsfile $(TargetName).deps.json $(HelixCorrelationPayloadPath)/xunit-runner/tools/netcoreapp2.0/xunit.console.dll $(TargetName).dll -notrait Category=SkipInCI -xml testResults.xml + $(HelixCorrelationPayloadPath)/xunit-runner/tools/net462/xunit.console.exe $(TargetName).dll -notrait Category=SkipInCI -xml testResults.xml + + + + _GetHelixWorkItemsOuter + _GetHelixWorkItemsInner + + + + + + + + + + + + + + + + + + $(TargetDir) + $(RunTestScript) + 01:00:00 + 00:30:00 + $(WorkItemTimeout) + + + + + + + /usr/local/opt/libomp/lib/libiomp5.dylib;/usr/local/opt/libomp/lib/libomp.dylib; + /usr/local/opt/libomp/lib/libomp.dylib; + + + + + + + + + + + + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/eng/testing/runTests.cmd b/eng/testing/runTests.cmd new file mode 100644 index 0000000000..818a6ff6e7 --- /dev/null +++ b/eng/testing/runTests.cmd @@ -0,0 +1,42 @@ +@echo off +setlocal ENABLEEXTENSIONS ENABLEDELAYEDEXPANSION + +set EXECUTION_DIR=%~dp0 + + +:: ========================= BEGIN Test Execution ============================= +echo ----- start %DATE% %TIME% =============== To repro directly: ===================================================== +echo pushd %EXECUTION_DIR% +echo ${runCommand} +echo popd +echo =========================================================================================================== +pushd %EXECUTION_DIR% +@echo on +${runCommand} +@set _exit_code=%ERRORLEVEL% +@echo off +if exist testResults.xml ( + set HAS_TEST_RESULTS=1 +) +popd +echo ----- end %DATE% %TIME% ----- exit code %_exit_code% ---------------------------------------------------------- +:: ========================= END Test Execution =============================== + +:: The tests either failed or crashed, copy output files +if not %_exit_code%==0 ( + if not "%HELIX_WORKITEM_UPLOAD_ROOT%" == "" ( + powershell Compress-Archive %EXECUTION_DIR%\TestOutput %HELIX_WORKITEM_UPLOAD_ROOT%\TestOutput.zip + ) +) + +:: The helix work item should not exit with non-zero if tests ran and produced results +:: The xunit console runner returns 1 when tests fail +if %_exit_code%==1 ( + if %HAS_TEST_RESULTS%==1 ( + if not "%HELIX_WORKITEM_PAYLOAD%"=="" ( + exit /b 0 + ) + ) +) + +exit /b %_exit_code% \ No newline at end of file diff --git a/eng/testing/runTests.sh b/eng/testing/runTests.sh new file mode 100644 index 0000000000..33a4276eff --- /dev/null +++ b/eng/testing/runTests.sh @@ -0,0 +1,38 @@ +#!/usr/bin/env bash + +EXECUTION_DIR=$(dirname "$0") + +# ========================= BEGIN Test Execution ============================= +echo ----- start $(date) =============== To repro directly: ===================================================== +echo pushd $EXECUTION_DIR +echo ${runCommand} +echo popd +echo =========================================================================================================== +pushd $EXECUTION_DIR +${runCommand} +test_exitcode=$? +if [[ -s testResults.xml ]]; then + has_test_results=1; +fi; +popd +echo ----- end $(date) ----- exit code $test_exitcode ---------------------------------------------------------- + +if [[ -n "${exitcode_list[$test_exitcode]}" ]]; then + echo exit code $test_exitcode means ${exitcode_list[$test_exitcode]} +fi +# ========================= END Test Execution =============================== + +# The tests either failed or crashed, copy output files +if [[ "$test_exitcode" != "0" && "$HELIX_WORKITEM_UPLOAD_ROOT" != ""]]; then + zip -vr $HELIX_WORKITEM_UPLOAD_ROOT/TestOutput.zip $EXECUTION_DIR/TestOutput/ +fi + +# The helix work item should not exit with non-zero if tests ran and produced results +# The xunit console runner returns 1 when tests fail +if [[ "$test_exitcode" == "1" && "$has_test_results" == "1" ]]; then + if [ -n "$HELIX_WORKITEM_PAYLOAD" ]; then + exit 0 + fi +fi + +exit $test_exitcode \ No newline at end of file diff --git a/test/Directory.Build.targets b/test/Directory.Build.targets index 6f4b2fc85e..60eff7f888 100644 --- a/test/Directory.Build.targets +++ b/test/Directory.Build.targets @@ -1,6 +1,7 @@ + - - - true - - true - - $HELIX_CORRELATION_PAYLOAD - %HELIX_CORRELATION_PAYLOAD% - - - - _GetHelixWorkItemsOuter - _GetHelixWorkItemsInner - - - - - - - - - - - - - - $(TargetDir) - dotnet exec --roll-forward Major --runtimeconfig $(TargetName).runtimeconfig.json --depsfile $(TargetName).deps.json $(HelixCorrelationPayloadPath)/xunit-runner/tools/netcoreapp2.0/xunit.console.dll $(TargetName).dll -notrait Category=SkipInCI -xml testResults.xml - $(HelixCorrelationPayloadPath)/xunit-runner/tools/net462/xunit.console.exe $(TargetName).dll -notrait Category=SkipInCI -xml testResults.xml - 01:00:00 - 00:30:00 - $(WorkItemTimeout) - - - - - - - /usr/local/opt/libomp/lib/libiomp5.dylib;/usr/local/opt/libomp/lib/libomp.dylib; - /usr/local/opt/libomp/lib/libomp.dylib; - - - - - - - - - - - - - - - - - - - - - - - - - - From fda37837e67fbbc532ddabb38e03f0899922f1ce Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Fri, 29 Mar 2024 06:10:38 -0700 Subject: [PATCH 23/26] Fix command to call script from helix --- eng/testing/helix.targets | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/eng/testing/helix.targets b/eng/testing/helix.targets index 9cd387278d..54b59b9dc7 100644 --- a/eng/testing/helix.targets +++ b/eng/testing/helix.targets @@ -10,7 +10,7 @@ %HELIX_CORRELATION_PAYLOAD% runTests.sh - runTests.cmd + ./runTests.cmd dotnet exec --roll-forward Major --runtimeconfig $(TargetName).runtimeconfig.json --depsfile $(TargetName).deps.json $(HelixCorrelationPayloadPath)/xunit-runner/tools/netcoreapp2.0/xunit.console.dll $(TargetName).dll -notrait Category=SkipInCI -xml testResults.xml $(HelixCorrelationPayloadPath)/xunit-runner/tools/net462/xunit.console.exe $(TargetName).dll -notrait Category=SkipInCI -xml testResults.xml @@ -44,7 +44,8 @@ $(TargetDir) - $(RunTestScript) + ./$(RunTestScript) + call $(RunTestScript) 01:00:00 00:30:00 $(WorkItemTimeout) From edcb44d48855dd94be33e89b348cb3abe4138635 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Fri, 29 Mar 2024 09:33:01 -0700 Subject: [PATCH 24/26] Only run .NETFramework tests on x64 Also cleanup setting RuntimeIdentifier since it's not required. Techincally we could be testing .NETFramework on more architectures by setting PlatformTarget but I don't want to open that can of worms yet since it might require new sets of baselines. --- eng/testing/helix.targets | 2 ++ test/Directory.Build.props | 1 - .../Microsoft.ML.PerformanceTests.csproj | 1 + 3 files changed, 3 insertions(+), 1 deletion(-) diff --git a/eng/testing/helix.targets b/eng/testing/helix.targets index 54b59b9dc7..8250dc5b8e 100644 --- a/eng/testing/helix.targets +++ b/eng/testing/helix.targets @@ -5,6 +5,8 @@ true true + + true $HELIX_CORRELATION_PAYLOAD %HELIX_CORRELATION_PAYLOAD% diff --git a/test/Directory.Build.props b/test/Directory.Build.props index 575d40302a..ac51c6ac78 100644 --- a/test/Directory.Build.props +++ b/test/Directory.Build.props @@ -4,7 +4,6 @@ $(NetMinimum);$(NetCurrent) $(TargetFrameworks);$(NetFrameworkMinimum) - win-x64 true false true diff --git a/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj b/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj index 9a625eb8fb..9c05539089 100644 --- a/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj +++ b/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj @@ -5,6 +5,7 @@ false false true + x64 true From 021c85dea81d9b47a8d51cdf4de4d2fb498d4b41 Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Fri, 29 Mar 2024 11:22:35 -0700 Subject: [PATCH 25/26] Fix typo in bash script --- eng/testing/runTests.sh | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/eng/testing/runTests.sh b/eng/testing/runTests.sh index 33a4276eff..02ed0d6dad 100644 --- a/eng/testing/runTests.sh +++ b/eng/testing/runTests.sh @@ -23,7 +23,7 @@ fi # ========================= END Test Execution =============================== # The tests either failed or crashed, copy output files -if [[ "$test_exitcode" != "0" && "$HELIX_WORKITEM_UPLOAD_ROOT" != ""]]; then +if [[ "$test_exitcode" != "0" && "$HELIX_WORKITEM_UPLOAD_ROOT" != "" ]]; then zip -vr $HELIX_WORKITEM_UPLOAD_ROOT/TestOutput.zip $EXECUTION_DIR/TestOutput/ fi From 1f7001146353aa4afd07ae3174e6de96daf8c4dc Mon Sep 17 00:00:00 2001 From: Eric StJohn Date: Fri, 29 Mar 2024 11:47:29 -0700 Subject: [PATCH 26/26] Build NETFramework projects for PlatformTarget that matches TargetArchitecture --- test/Directory.Build.props | 1 + .../Microsoft.ML.PerformanceTests.csproj | 3 ++- 2 files changed, 3 insertions(+), 1 deletion(-) diff --git a/test/Directory.Build.props b/test/Directory.Build.props index ac51c6ac78..61418b5fc9 100644 --- a/test/Directory.Build.props +++ b/test/Directory.Build.props @@ -4,6 +4,7 @@ $(NetMinimum);$(NetCurrent) $(TargetFrameworks);$(NetFrameworkMinimum) + $(TargetArchitecture) true false true diff --git a/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj b/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj index 9c05539089..dc284fcdbb 100644 --- a/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj +++ b/test/Microsoft.ML.PerformanceTests/Microsoft.ML.PerformanceTests.csproj @@ -5,8 +5,9 @@ false false true - x64 true + + false