Craig R

My feedback

  1. 8 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Data Factory  ·  Flag idea as inappropriate…  ·  Admin →
    Craig R shared this idea  · 
  2. 5,668 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    223 comments  ·  Storage » Files  ·  Flag idea as inappropriate…  ·  Admin →

    Thanks for the feedback! We are interested in collecting feedback on this request – please vote for it if this is something you like to see.

    We’re also interested in learning more what people want to use the SFTP/FTPS for and which protocol they prefer. Please feel free to leave us a comment letting us know more detail!

    Thanks,

    Will Gries
    Program Manager, Azure Files

    Craig R supported this idea  · 
    An error occurred while saving the comment
    Craig R commented  · 

    Please add this feature. A good portion of companies are still stuck in SFTP land due to software and skillset limitations (i.e. they are not scripting or command-line savvy).

    Furthermore, with SFTP support, we will be able to leverage the new Azure Data Factory v2 Event-based Triggers without "yet another complex hacky azure integration + vm + logic app + whatever".

  3. 34 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Data Factory  ·  Flag idea as inappropriate…  ·  Admin →
    Craig R shared this idea  · 
  4. 118 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    7 comments  ·  Data Lake  ·  Flag idea as inappropriate…  ·  Admin →
    An error occurred while saving the comment
    Craig R commented  · 

    One challenge with the sample json extractor is that it does not scale and only handles arrays of json, while the majority of parsers support jsonlines, or sets of json objects.

    The scaling issue is similar to early XML parsers that chose to load the DOM into memory completely vs. streaming content (e.g. SAX) for scalability.
    --> to expose the limitation, produce some array of json objects that is sufficiently large (100G works, but probably smaller works also) and you'll see one vertex allocated and your job will hang forever... no idea what it is doing :)

    ADLA does not handle binary data well at all , and really doesn't have an auto-scale capability for input and output, imho. For example, the read is generally one vertex / AU per file, and many of these files are one file.

    To handle this, btw, the guidance is to use ADF to split the file into "many" small input files, then ADLA will scale. (or, switch to spark and hack it out)

    Craig R supported this idea  · 
  5. 112 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    under review  ·  5 comments  ·  Azure Synapse Analytics » SQL/T-SQL  ·  Flag idea as inappropriate…  ·  Admin →
    Craig R supported this idea  · 
  6. 8 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Data Factory  ·  Flag idea as inappropriate…  ·  Admin →
    Craig R supported this idea  · 

Feedback and Knowledge Base