Be sure to review our Idea Submission Guidelines for more information!
Submission GuidelinesHello,
After used the new "Image Recognition Tool" a few days, I think you could improve it :
> by adding the dimensional constraints in front of each of the pre-trained models,
> by adding a true tool to divide the training data correctly (in order to have an equivalent number of images for each of the labels)
> at least, allow the tool to use black & white images (I wanted to test it on the MNIST, but the tool tells me that it necessarily needs RGB images) ?
Question : do you in the future allow the user to choose between CPU or GPU usage ?
In any case, thank you again for this new tool, it is certainly perfectible, but very simple to use, and I sincerely think that it will allow a greater number of people to understand the many use cases made possible thanks to image recognition.
Thank you again
Kévin VANCAPPEL (France ;-))
Thank you again.
Kévin VANCAPPEL
Our DAT file structure is as follows:
The first line of the .DAT file must be a header row identifying the field neames.
The .DAT file must use the following Concordance default delimiters:
Comma ASCII character (020)
Quote ASCII character (254)
Newline ASCII character (174)
Thank you,
Pete Vara
Hi,
I've been working on reporting for a while now and figure out that creatitng sub total wasn't part of any tool.
Any chance this could be implemented in next versions or any macro available?
Thanks
Simon
I constantly find my using pre and post SQL Commands in the Output tool to run SQL when I don't actually have any data to output.
One example is when I load data into S3 and want to load it into Redshift. I have SQL code to run but no data to Output - I end up running a dummy row into a temp table.
So can we have an SQL tool that simply acts the same as a Pre-SQL command without the associated data output. Once the command is run we should be able to continue the workflow, so the tool should have an option input and output, like the Run Command tool.
Preface: I have only used the in-DB tools with Teradata so I am unsure if this applies to other supported databases.
When building a fairly sophisticated workflow using in-DB tools, sometimes the workflow may fail due to the underlying queries running up against CPU / Memory limits. This is most common when doing several joins back to back as Alteryx sends this as one big query with various nested sub queries. When working with datasets in the hundereds of millions and billions of records, this can be extremely taxing for the DB to run as one huge query. (It is possible to get arround this by using in-DB write out to a temporary table as an intermediate step in the workflow)
When a routine does hit a in-DB resource limit and the DB kills the query, it causes Alteryx to immediately fail the workflow run. Any "temporary" tables Alteryx creates are in reality perm tables that Alteryx usually just drops at the end of a successful run. If the run does not end successfully due to hitting a resource limit, these "Temporary" (perm) tables are not dropped. I only noticed this after building out a workflow and running up against a few resource limits, I then started getting database out of space errors. Upon looking into it, I found all the previously created "temporary" tables were still there and taking up many TBs of space.
My proposed solution is for Alteryx's in-DB tools to drop any "temporary" tables it has created when a run ends - regardless of if the entire module finished successfully.
Thanks,
Ryan
When a workflow has run it has it's final message.
It would be great to have a datetime stamp included in the file message. When running multiple jobs and editing multiple workflows it would be great to be able to check the last runtime so you know when each job had finished. As I often flip between multiple workflows during development and can working on up to 10 or more at any one time
Hi,
I think that the sample tool should have a T or F port.
Lets say I want to keep first N records but would like to stream the rest of the data (the not sampled one) somewhere else in my workflow, its possible but it would be easier to have that in the sampler.
Simon
Korem
When it comes to something going wrong in Alteryx the last thing I can ever remember is the email address to use for support. Instead of trying to remember whether to use customersuccess@ or clientsvcs@ why can't it just simply be support@?
At your time of need, please make it as simple as possible to get help.
When using Server datasources Alteryx can take a long time to query metadata, particularly for long complex queries. This often happens on opening a module or clicking on the Input tool to eit properties.
This leads to frustration with these modules. It would be good for Alteryx to cache the metadata (i.e. columns) from these inputs and prompt the user to reuse this cached data if it it takes longer than say 2 seconds to retrieve the query.
As you may be aware localisation is the adapting of computer software to regional differences of a target market, Internationalization is the process of designing software so that it can potentially be adapted to various languages without engineering changes.
The idea is to make Alteryx designer tool, the web help content and example workflows to be multilingual (Possibly with the use of "lic" language files or similar) Hopefully the sotware and tutorials will all be localised by crowdsourcing initiatives within the Alteryx community.
I sincerely believe this will help the tool get a lot of traction not in US and UK but in other parts of the world,
Highly likely Mandarin and Spanish would be the first two language versions...
Top languages by population per Nationalencyklopedin 2007 (2010)Language Native speakers(millions) Fraction of worldpopulation (2007)
Mandarin | 935 (955) | 14.1% |
Spanish | 390 (405) | 5.85% |
English | 365 (360) | 5.52% |
Hindi | 295 (310)[2] | 4.46% |
Arabic | 280 (295) | 4.43% |
Portuguese | 205 (215) | 3.08% |
Bengali | 200 (205) | 3.05% |
Russian | 160 (155) | 2.42% |
Japanese | 125 (125) | 1.92% |
Punjabi | 95 (100) | 1.44% |
I am trying to run batch regressions on a pretty sizable set of data. About ~1M distinct groups of data, each wtih 30-500 x,y pairs.
A batch macro with a linear regression works ok - but it is really slow. Started at about 2-3s per regression. After stripping out bunch or reporting from the macro, I am down to ~2s. This is still feels quite slow compared to something purpose built.
Has anyone experimented with higher speed versions that just dump out m,b, & r2?
Hi - I miss the functionality in 9.5 of being able to set a default tool in a tab and then drag in tools from the tab. This seems to be gone in 10. Is there any possibility of it coming back?
Many thanks - Nathalie
Hi,
I've been using desktop scheduler to download and parse out a streaming JSON file. My script takes 4 seconds to execute, the data is updated on a per second basis.
Currently, my only option is to execute the scheduled job at the lowest level of granularity in the temporal at 1 minute. therefore I'm missing records.
Would it be possible to add a second(s) option to the scheduler?
I can see particular benefits from doing a CRON at under a minute, especially with event data capture.
Best Regards,
Allan
Currently only DateTime based functions are available, Time based functions should be introduced. like TimeAdd(), TimeDiff() etc.
This will help users a lot to calculate different aspects of time based calculations...
Ashok Bhatt
Given redshift prefers accepting many small files for bulk loading into redshift, it would be good to be able to have a max record limit within the s3 upload tool (similar to functionality for s3 download)
The other functionality that is useful for the s3 upload tool is ability to append file names based on datetimestamp_001, 002, 003 etc similar to current output tool
Within the select tool when you have hilighted a set of rows it would be NICE to be able to RIGHT-CLICK for OPTIONS rather than have to move the cursor up to the options to get to choices.
Thanks,
Mark
Hi,
Carlson Companies is moving to a Vertica environment and it would be great if that was supported with the In-database tools. That would definitely help and expand the use of Alteryx at our company!
Thanks,
Tyler Mittelstadt
At the moment, we are not able to use input data field names and its values in Output tool, mainly in the Pre-SQL and Post-SQL statement. I see some discussions on this in the community and in many scenarios we require that. It will be great if we have this option.
Idea:
A funcionality added to the Impute values tool for multiple imputation and maximum likelihood imputation of fields with missing at random will be very useful.
Rationale:
Missing data form a problem and advanced techniques are complicated. One great idea in statistics is multiple imputation,
filling the gaps in the data not with average, median, mode or user defined static values but instead with plausible values considering other fields.
SAS has PROC MI tool, here is a page detailing the usage with examples: http://www.ats.ucla.edu/stat/sas/seminars/missing_data/mi_new_1.htm
Also there is PROC CALIS for maximum likelihood here...
Same useful tool exists in spss as well http://www.appliedmissingdata.com/spss-multiple-imputation.pdf
Best
I've come to realize that the JOIN tool is case-sensitive by design but it would be helpful if you could turn that behavior on/off (via checkbox?) within the JOIN tool. For those of us that work predominantly in database environments that are not case-sensitive, this default behavior has caused me problems many times. Having to force the case to either upper or lower upstream of the JOIN on both flows in order to ensure a successful join is an extra step that would not be necessary if you could disable case-sensitive with a checkbox.
There is a web hosted trial that anyone can have a hands on experiance with alteryx tutorials without even downoading the tool.
That's awesome... http://goo.gl/dpSoe2
It may be a nice idea to;
1) either start seperate "Alteryx-kaggle" instances with data sets specific to each kaggle competition so that anyone want to try out may have a go with those well known examples thru the Alteryx site,
2) Or even better have a partnership with kaggle so that anyone can just have it's own Alteryx trial per specific competition on the kaggle website...
I'm sure this will draw a lot of attention...
Rationale;
You'll immediately have a greater reach in Kaggle community, some data hobbiyists and cs, ie students and acedemics (which will eventually end up doing lot's of data blending when ther are going to be hired by top notch firms...
User | Likes Count |
---|---|
10 | |
7 | |
5 | |
5 | |
3 |