I need to import 1000+ Excel files, some of our customers submit data which can earn them rebate. We have a desired template but over the last few years this has not been enforced and I now need to import the data for 2 reasons, the first to monitor who is submitting data & assessing the format, the second to store the data for reporting purposes.
I have a flow with a dynamic Directory connection and two macros, one to find all files in the folder, the other to find all sheet names within the files to bring all sheets into my flow.
Where I'm really having problems though is that the initial dataset is one huge dump of all files and every combination of headings, I can't seem to separate out which headings are in which file so that I can assess them for the correct format and build a usable dataset, I have over 380 fields because of this variation over the years. Is there a way of building a view to show each individual file, each sheet name within the file and the contents of the first 5 rows?
I've attached sample data to show some of the variety, note the different sheets are all in different files from different customers.
There's a lot going on here. So I'll start with some general recommendations:
Overall, though, with so many files, it's not so much a matter of bringing it all in. I'd be curious what you can even do to clean 'em up after you figure out how to get all those variations in your workflow... I think your best bet is to hedge the incentive of the rebate and force conformity towards an acceptable data profile and go from there. Only due to the volume of sheets.
Hope that helps, -Jay

