Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature request/question on EDR support for FS. #154

Open
tmefford opened this issue Jan 30, 2020 · 7 comments
Open

Feature request/question on EDR support for FS. #154

tmefford opened this issue Jan 30, 2020 · 7 comments

Comments

@tmefford
Copy link

The beegfs example doesn't use the IB interface. The lustre example doesn't seem to either. How could I get that to work for either? I'd have to use different nodes for the beegfssm, at least. Failing that, what is the best way to get I/O performance? Just bump the number of instances of beegfssm? Sorry for being a little open-ended with the question. Suggestions appreciated.

Thank you

@xpillons
Copy link
Collaborator

the only VM supporting IB are our HPC skus which are not design for high density storage. If you want, you can use these for lustre so that your compute and storage will be on the same IB fabric. What level of I/O performance are you looking for ?

@garvct
Copy link
Collaborator

garvct commented Jan 31, 2020

Please consult the following blog post for tips on tuning beeGFS/BeeOND on Azure.
https://techcommunity.microsoft.com/t5/azurecat/tuning-beegfs-and-beeond-on-azure-for-specific-i-o-patterns/ba-p/1015446

@tmefford
Copy link
Author

Thanks for the replies. I'm not an I/O expert. I don't have a quantifiable performance target, I only know that even single node runs are way worse on the BeeGFS filesystem than on local SSD. Using an HPC SKU for lustre seems to make sense, but I don't know what i would need to do to make it use the IB interface. If I don't go the IB route, what would you expect the best performing azurehpc filesystem to be? lustre_full with maybe an increased number of instances? I will look at the BeeGFS docs, but, even though ideally I would, I don't really know the apps' I/O pattern. Thank you for your time, suggestions appreciated.

@edwardsp
Copy link
Collaborator

edwardsp commented Feb 4, 2020

You can actually use the local disks in the HPC VMs but it will probably be a fairly expensive way to create a parallel filesystem (since the disks are only 700GiB per VM). You would just need to make sure you use /dev/sdb when calling the lfsmaster and lfsoss scripts in the install section. This would then allow you to use the IB network. But, if you want compute on that same network you will need to provision more VMs in the scaleset and set the others up as compute (only VMs within a scaleset can communicate with one-another on the IB network).

You should only be limited by the network throughput using the Lv2 setup.

Can you provide more details about what you are running?

@tmefford
Copy link
Author

tmefford commented Feb 4, 2020 via email

@tmefford
Copy link
Author

tmefford commented Feb 4, 2020

Super simple question--How would I alter this to get two non-OS SSDs per compute node?

    "compute": {
        "type": "vmss",
        "vm_type": "Standard_HB60rs",
        "accelerated_networking": false,
        "instances": 2,
        "image": "variables.hpcimage",
        "subnet": "compute",
        "tags": [
            "disable-selinux",
            "lfsrepo",
            "lfsclient",
            "localuser",
            "pbsclient",
            "nfsclient"
        ]
    },

Thanks,

Tim

@edwardsp
Copy link
Collaborator

edwardsp commented Feb 5, 2020

Hi Tim,

You can just add:

    "storage_sku": "Premium_LRS",
      "data_disks": [
        4095, 4095
      ],

The "storage_sku" is the type of storage and the "data_disks" are a list of disks to add (if disk size is less than 4096 it will use caching provided the VM type supports it). For the above it is premium SSD. But, please rememeber, these SSDs are not inside the physical VM. There is still a latency.

Best regards,
Paul

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants