-
Notifications
You must be signed in to change notification settings - Fork 1.9k
Adding Multiple Training Files to the Pipeline? #192
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Thanks for asking! This is not currently possible, but let's use this issue to track enabling multiple inputs in a pipeline. Just to clarify: is your intention to concatenate the two files as soon as they are loaded, or to apply different transforms/trainers to them? A potential workaround for now is to read in the examples from both files into memory and use the |
My intention is for creating and testing ML structures with large datasets to be modular and less taxing on file transfers to/from servers. For example, moving 100GB is to a server is easier if split by time or another parameter. It also allows ML structures to be updated as new data comes in without having to concat onto what are already are/is a large file. Reducing the memory footprint by loading subsets of the data would be nice, but as I understand it, that is not possible for all ML structures. I have concated the files and it works properly but this would be a nice feature to have. Thanks for the answer. |
@cflint987, |
DRI RESPONSE: You can do it with new Api:
Please let me know if this satisfy you. I'm intend to close this issue within next few days. |
System information
OS version/distro: Windows 7 Home
.NET Version (eg., dotnet --info): ML .net V0.1.0
Issue:
What is the correct way to add multiple training files to a Learning Pipeline?
In the Taxi Fare example, just adding another textloader and/or ColumnCopier, etc seems to not be correct.
Example:
pipeline.Add(new TextLoader(DataPath, useHeader: true, separator: ","));
pipeline.Add(new TextLoader(DataPath2, useHeader: true, separator: ","));
The text was updated successfully, but these errors were encountered: